Web spider framework that can spider a domain and collect pages it visits.
~~~bash $ apt install openssl libssl-dev ~~~
~~~bash $ git clone https://github.com/madeindjs/spider.git $ cd spider $ cargo build --release $ ./target/debug/rust-crawler http://localhost:4000 ~~~
Will produce something like this
- http://localhost:4000/
- http://localhost:4000/portfolio
- http://localhost:4000/resume
- http://localhost:4000/blog
Add this dependency to your Cargo.toml file.
~~~toml [dependencies] TODO ~~~
~~~rust // insert crate here
let mut localhost = Website::new("http://localhost:4000"); localhost.crawl();
for page in localhost.getpages() { println!("- {}", page.geturl()); }
// TODO ~~~
I am open-minded to any contribution. Just fork & commit
on another branch.