⚡ A lightning-fast web-crawler which collects links of a given host recursively and categorizes them to internal and external links.
- Parallel work get queries faster with parallel works when websites have much more pages to request
- RESTful API run the server in background and integrate it in your technical stack
For getting started you will need rust and cargo installed on your computer. Then follow commands below:
To run
cargo run
To build:
cargo build --release
after building, you will have binary file located in target/release/link_collector
, then just execute it.
Server address: http://0.0.0.0:4000
Available requests
http://0.0.0.0:4000/links?url=https://my.website.com
Query url
takes seed host to look up links for.
After running server, request to http://0.0.0.0:4000/links?url=https://www.rust-lang.org
to get all links from official rust programming language. Result below:
LinkCollector is open-source, so if you have any idea to improve or want to contribute feel free to open pull requests!