Cleaned up README

It's an attempt, at least.
This commit is contained in:
Felipe 2025-01-23 15:27:33 -03:00 committed by GitHub
parent 60e639c15a
commit 537656053f
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

268
README.md
View File

@ -6,142 +6,128 @@ This is a fork of the [Wayback Machine Downloader](https://github.com/hartator/w
Included here is partial content from other forks, namely those @ [ShiftaDeband](https://github.com/ShiftaDeband/wayback-machine-downloader) and [matthid](https://github.com/matthid/wayback-machine-downloader) — attributions are in the code and go to the original authors; as well as a few additional (future) features. Included here is partial content from other forks, namely those @ [ShiftaDeband](https://github.com/ShiftaDeband/wayback-machine-downloader) and [matthid](https://github.com/matthid/wayback-machine-downloader) — attributions are in the code and go to the original authors; as well as a few additional (future) features.
## Installation ## ▶️ Quick start
Note: You need to install Ruby on your system (>= 2.3) to run this program — if you don't already have it.
1. Clone/download this repository Download a website's latest snapshot:
2. In your terminal (e.g. Command Prompt, PowerShell, Windows Terminal), navigate to the directory where you cloned/downloaded this repository
3. Navigate to `wayback_machine_downloader\bin` (psst, Windows users: open this directory in File Explorer, then press Shift + Right Click → "Open Terminal here")
4. Run:
```bash ```bash
ruby wayback_machine_downloader [options] URL ruby wayback_machine_downloader https://example.com
```
Your files will save to `./websites/example.com/` with their original structure preserved.
## 📥 Installation
### Requirements
- Ruby 2.3+ ([download Ruby here](https://www.ruby-lang.org/en/downloads/))
- Bundler gem (`gem install bundler`)
### Step-by-step setup
1. **Install Ruby**:
```bash
ruby -v
```
This will verify your installation. If not installed, [download Ruby](https://www.ruby-lang.org/en/downloads/) for your OS.
2. **Install dependencies**:
```bash
bundle install
``` ```
### Using Docker 3. **Run it**:
```bash
cd path/to/wayback-machine-downloader/bin
ruby wayback_machine_downloader https://example.com
```
For example, if you extracted the contents to a folder named "wayback-machine-downloader" in your Downloads directory, you'd need to type `cd Downloads\wayback-machine-downloader\bin`.
*Windows tip*: In File Explorer, Shift + Right Click your `bin` folder → "Open Terminal here".
## 🐳 Docker users
We have a Docker image! See [#Packages](https://github.com/StrawberryMaster/wayback-machine-downloader/pkgs/container/wayback-machine-downloader) for the latest version. You can also build it yourself. Here's how: We have a Docker image! See [#Packages](https://github.com/StrawberryMaster/wayback-machine-downloader/pkgs/container/wayback-machine-downloader) for the latest version. You can also build it yourself. Here's how:
```bash ```bash
docker build -t wayback_machine_downloader . docker build -t wayback_machine_downloader .
docker run -it --rm wayback_machine_downloader [options] URL docker run -it --rm wayback_machine_downloader [options] URL
``` ```
# Constants ## ⚙️ Configuration
There are a few constants that can be edited in the `wayback_machine_downloader.rb` file for your convenience. The default values may be conservative, so you can adjust them to your needs. They are: There are a few constants that can be edited in the `wayback_machine_downloader.rb` file for your convenience. The default values may be conservative, so you can adjust them to your needs. They are:
- `DEFAULT_TIMEOUT` - The default timeout (in seconds) for HTTP requests. Default is 30 seconds. ```ruby
- `MAX_RETRIES` - The maximum number of retries for HTTP requests. Default is 3. DEFAULT_TIMEOUT = 30 # HTTP timeout (in seconds)
- `RETRY_DELAY` - The delay (in seconds) between retries for HTTP requests. Default is 2 seconds. MAX_RETRIES = 3 # Failed request retries
- `RATE_LIMIT` - The rate limit (in seconds) for HTTP requests. Default is 0.25 seconds. RETRY_DELAY = 2 # Wait between retries
- `CONNECTION_POOL_SIZE` - The size of the HTTP connection pool. Default is 10 connections. RATE_LIMIT = 0.25 # Throttle between requests
- `MEMORY_BUFFER_SIZE` - The size of the memory buffer (in bytes) for downloads. Default is 16KB. CONNECTION_POOL_SIZE = 10 # No. of simultaneous connections
MEMORY_BUFFER_SIZE = 16384 # Size of download buffer
---
## Instructions
### Basic usage
Run `wayback_machine_downloader` with the base URL of the website you want to retrieve as a parameter (e.g., https://example.com):
```bash
ruby wayback_machine_downloader https://example.com
``` ```
## How it works ## 🛠️ Advanced usage
It will download the last version of every file present on Wayback Machine to `./websites/example.com/`. It will also re-create a directory structure and auto-create `index.html` pages to work seamlessly with Apache and Nginx. All files downloaded are the original ones and not Wayback Machine rewritten versions; this way, URLs and links structure are the same as before.
## Advanced usage ### Basic options
``` | Option | Description |
Usage: ruby wayback_machine_downloader https://example.com |--------|-------------|
| `-d DIR`, `--directory DIR` | Custom output directory |
| `-s`, `--all-timestamps` | Download all historical versions |
| `-f TS`, `--from TS` | Start from timestamp (e.g., 20060121) |
| `-t TS`, `--to TS` | Stop at timestamp |
| `-e`, `--exact-url` | Download exact URL only |
| `-r`, `--rewritten` | Download rewritten Wayback Archive files only |
Download an entire website from the Wayback Machine. **Example** - Download files to `downloaded-backup` folder
Optional options:
-d, --directory PATH Directory to save the downloaded files into
Default is ./websites/ plus the domain name
-s, --all-timestamps Download all snapshots/timestamps for a given website
-f, --from TIMESTAMP Only files on or after timestamp supplied (ie. 20060716231334)
-t, --to TIMESTAMP Only files on or before timestamp supplied (ie. 20100916231334)
-e, --exact-url Download only the url provided and not the full site
-o, --only ONLY_FILTER Restrict downloading to urls that match this filter
(use // notation for the filter to be treated as a regex)
-x, --exclude EXCLUDE_FILTER Skip downloading of urls that match this filter
(use // notation for the filter to be treated as a regex)
-a, --all Expand downloading to error files (40x and 50x) and redirections (30x)
-c, --concurrency NUMBER Number of multiple files to download at a time
Default is one file at a time (ie. 20)
-p, --maximum-snapshot NUMBER Maximum snapshot pages to consider (Default is 100)
Count an average of 150,000 snapshots per page
-l, --list Only list file urls in a JSON format with the archived timestamps, won't download anything
```
### Specify directory to save files to
-d, --directory PATH
Optional. By default, Wayback Machine Downloader will download files to `./websites/` followed by the domain name of the website. You may want to save files in a specific directory using this option.
Example:
```bash ```bash
ruby wayback_machine_downloader https://example.com --directory downloaded-backup/ ruby wayback_machine_downloader https://example.com --directory downloaded-backup/
``` ```
By default, Wayback Machine Downloader will download files to ./websites/ followed by the domain name of the website. You may want to save files in a specific directory using this option.
### All timestamps **Example 2** - Download historical timestamps:
-s, --all-timestamps
Optional. This option will download all timestamps/snapshots for a given website. It will uses the timestamp of each snapshot as directory.
Example:
```bash ```bash
ruby wayback_machine_downloader https://example.com --all-timestamps ruby wayback_machine_downloader https://example.com --all-timestamps
```
Will download: This option will download all timestamps/snapshots for a given website. It will uses the timestamp of each snapshot as directory. In this case, it will download, for example:
```bash
websites/example.com/20060715085250/index.html websites/example.com/20060715085250/index.html
websites/example.com/20051120005053/index.html websites/example.com/20051120005053/index.html
websites/example.com/20060111095815/img/logo.png websites/example.com/20060111095815/img/logo.png
... ...
``` ```
### From timestamp **Example 3** - Download content on or after July 16, 2006:
-f, --from TIMESTAMP
Optional. You may want to supply a from timestamp to lock your backup to a specific version of the website. Timestamps can be found inside the urls of the regular Wayback Machine website (e.g., https://web.archive.org/web/20060716231334/http://example.com). You can also use years (2006), years + month (200607), etc. It can be used in combination of To Timestamp.
Wayback Machine Downloader will then fetch only file versions on or after the timestamp specified.
Example:
```bash ```bash
ruby wayback_machine_downloader https://example.com --from 20060716231334 ruby wayback_machine_downloader https://example.com --from 20060716231334
``` ```
You may want to supply a from timestamp to lock your backup to a specific version of the website. Timestamps can be found inside the urls of the regular Wayback Machine website (e.g., https://web.archive.org/web/20060716231334/http://example.com). You can also use years (2006), years + month (200607), etc. It can be used in combination of To Timestamp.
Wayback Machine Downloader will then fetch only file versions on or after the timestamp specified.
### To timestamp **Example 4** - Download content on or before September 16, 2010:
```bash
-t, --to TIMESTAMP ruby wayback_machine_downloader https://example.com --to 20100916231334
```
Optional. You may want to supply a to timestamp to lock your backup to a specific version of the website. Timestamps can be found inside the urls of the regular Wayback Machine website (e.g., https://web.archive.org/web/20100916231334/http://example.com). You can also use years (2010), years + month (201009), etc. It can be used in combination of From Timestamp. You may want to supply a to timestamp to lock your backup to a specific version of the website. Timestamps can be found inside the urls of the regular Wayback Machine website (e.g., https://web.archive.org/web/20100916231334/http://example.com). You can also use years (2010), years + month (201009), etc. It can be used in combination of From Timestamp.
Wayback Machine Downloader will then fetch only file versions on or before the timestamp specified. Wayback Machine Downloader will then fetch only file versions on or before the timestamp specified.
Example: **Example 5** - Download the homepage of http://example.com
ruby wayback_machine_downloader https://example.com --to 20100916231334
### Exact url
-e, --exact-url
Optional. If you want to retrieve only the file matching exactly the url provided, you can use this flag. It will avoid downloading anything else.
For example, if you only want to download only the html homepage file of example.com:
```bash ```bash
ruby wayback_machine_downloader https://example.com --exact-url ruby wayback_machine_downloader https://example.com --exact-url
``` ```
If you want to retrieve only the file matching exactly the url provided, you can use this flag. It will avoid downloading anything else.
### Only URL filter **Example 6** - Download a rewritten file
```bash
ruby wayback_machine_downloader https://example.com --rewritten
```
Useful if you want to download the rewritten files from the Wayback Machine instead of the original ones.
-o, --only ONLY_FILTER ### Filtering Content
| Option | Description |
|--------|-------------|
| `-o FILTER`, `--only FILTER` | Only download matching URLs (supports regex) |
| `-x FILTER`, `--exclude FILTER` | Exclude matching URLs |
Optional. You may want to retrieve files which are of a certain type (e.g., .pdf, .jpg, .wrd...) or are in a specific directory. To do so, you can supply the `--only` flag with a string or a regex (using the '/regex/' notation) to limit which files Wayback Machine Downloader will download. **Example** - Include only images:
```bash
For example, if you only want to download files inside a specific `my_directory`: ruby wayback_machine_downloader https://example.com -o "/\.(jpg|png)/i"
```
You may want to retrieve files which are of a certain type (e.g., .pdf, .jpg, .wrd...) or are in a specific directory. To do so, you can supply the --only flag with a string or a regex (using the '/regex/' notation) to limit which files Wayback Machine Downloader will download.
For example, if you only want to download files inside a specific my_directory:
```bash ```bash
ruby wayback_machine_downloader https://example.com --only my_directory ruby wayback_machine_downloader https://example.com --only my_directory
``` ```
@ -150,13 +136,12 @@ Or if you want to download every images without anything else:
ruby wayback_machine_downloader https://example.com --only "/\.(gif|jpg|jpeg)$/i" ruby wayback_machine_downloader https://example.com --only "/\.(gif|jpg|jpeg)$/i"
``` ```
### Exclude URL filter **Example 2** - Exclude images:
```bash
-x, --exclude EXCLUDE_FILTER ruby wayback_machine_downloader https://example.com -x "/\.(jpg|png)/i"
```
Optional. You may want to retrieve files which aren't of a certain type (e.g., .pdf, .jpg, .wrd...) or aren't in a specific directory. To do so, you can supply the `--exclude` flag with a string or a regex (using the '/regex/' notation) to limit which files Wayback Machine Downloader will download. You may want to retrieve files which aren't of a certain type (e.g., .pdf, .jpg, .wrd...) or aren't in a specific directory. To do so, you can supply the --exclude flag with a string or a regex (using the '/regex/' notation) to limit which files Wayback Machine Downloader will download.
For example, if you want to avoid downloading files inside my_directory:
For example, if you want to avoid downloading files inside `my_directory`:
```bash ```bash
ruby wayback_machine_downloader https://example.com --exclude my_directory ruby wayback_machine_downloader https://example.com --exclude my_directory
``` ```
@ -165,55 +150,48 @@ Or if you want to download everything except images:
ruby wayback_machine_downloader https://example.com --exclude "/\.(gif|jpg|jpeg)$/i" ruby wayback_machine_downloader https://example.com --exclude "/\.(gif|jpg|jpeg)$/i"
``` ```
### Expand downloading to all file types ### Performance
| Option | Description |
|--------|-------------|
| `-c NUM`, `--concurrency NUM` | Concurrent downloads (default: 1) |
| `-p NUM`, `--maximum-snapshot NUM` | Max snapshot pages (150k snapshots/page) |
-a, --all **Example** - 20 parallel downloads:
Optional. By default, Wayback Machine Downloader limits itself to files that responded with 200 OK code. If you also need errors files (40x and 50x codes) or redirections files (30x codes), you can use the `--all` or `-a` flag and Wayback Machine Downloader will download them in addition of the 200 OK files. It will also keep empty files that are removed by default.
Example:
```bash
ruby wayback_machine_downloader https://example.com --all
```
### Only list files without downloading
-l, --list
It will just display the files to be downloaded with their snapshot timestamps and urls. The output format is JSON. It won't download anything. It's useful for debugging or to connect to another application.
Example:
```bash
ruby wayback_machine_downloader https://example.com --list
```
### Maximum number of snapshot pages to consider
-p, --snapshot-pages NUMBER
Optional. Specify the maximum number of snapshot pages to consider. Count an average of 150,000 snapshots per page. 100 is the default maximum number of snapshot pages and should be sufficient for most websites. Use a bigger number if you want to download a very large website.
Example:
```bash
ruby wayback_machine_downloader https://example.com --snapshot-pages 300
```
### Download multiple files at a time
-c, --concurrency NUMBER
Optional. Specify the number of multiple files you want to download at the same time. Allows one to speed up the download of a website significantly. Default is to download one file at a time.
Example:
```bash ```bash
ruby wayback_machine_downloader https://example.com --concurrency 20 ruby wayback_machine_downloader https://example.com --concurrency 20
``` ```
Will specify the number of multiple files you want to download at the same time. Allows one to speed up the download of a website significantly. Default is to download one file at a time.
## Contributing **Example 2** - 300 snapshot pages:
```bash
ruby wayback_machine_downloader https://example.com --snapshot-pages 300
```
Will specify the maximum number of snapshot pages to consider. Count an average of 150,000 snapshots per page. 100 is the default maximum number of snapshot pages and should be sufficient for most websites. Use a bigger number if you want to download a very large website.
Contributions are welcome! Just submit a pull request via GitHub. ### Diagnostics
| Option | Description |
|--------|-------------|
| `-a`, `--all` | Include error pages (40x/50x) |
| `-l`, `--list` | List files without downloading |
To run the tests: **Example** - Download all files
```bash
ruby wayback_machine_downloader https://example.com --all
```
By default, Wayback Machine Downloader limits itself to files that responded with 200 OK code. If you also need errors files (40x and 50x codes) or redirections files (30x codes), you can use the --all or -a flag and Wayback Machine Downloader will download them in addition of the 200 OK files. It will also keep empty files that are removed by default.
bundle install **Example 2** - Generate URL list:
```bash
ruby wayback_machine_downloader https://example.com --list
```
It will just display the files to be downloaded with their snapshot timestamps and urls. The output format is JSON. It won't download anything. It's useful for debugging or to connect to another application.
## 🤝 Contributing
1. Fork the repository
2. Create a feature branch
3. Submit a pull request
**Run tests** (note, these are still broken!):
```bash
bundle exec rake test bundle exec rake test
```