๐ท๏ธ WebCrawlerX ๐
A flexible and efficient web crawler written in Rust.
- Multiple spider implementations (CVE Details, GitHub, Quotes)
- Configurable crawling parameters (delay, concurrent requests, page limit)
- Easy to extend with new spiders
cargo install webcrawlerx
List available spiders:
webcrawlerx spiders
Run a specific spider:
webcrawlerx run --spider <spider_name>
--spider <spider_name> [--delay <ms>] [--concurrent <num>] [--limit <num>]
Example:
webcrawlerx run --spider cvedetails --delay 200 --concurrent 2 --limit 10
To add a new spider, create a new module in the spiders
directory and implement the Spider
trait. Then, update the run_spider
function in main.rs
to include your new spider.
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.