Web Crawler

Finds every page, image, and script on a website (and downloads it)

Usage

``` Rust Web Crawler

Usage: web-crawler [OPTIONS]

Arguments:

Options: -d, --download Download all files -c, --crawl-external Whether or not to crawl other websites it finds a link to. Might result in downloading the entire internet -m, --max-url-length Maximum url length it allows. Will ignore page it url length reaches this limit [default: 300] -e, --exclude Will ignore paths that start with these strings (comma-seperated) --export Where to export found URLs --export-internal Where to export internal URLs --export-external Where to export external URLs -t, --timeout Timeout between requests in milliseconds [default: 100] -h, --help Print help -V, --version Print version ```

How to compile yourself

  1. Download rust
  2. Type cargo build -r
  3. Executable is in target/release