๐ท๏ธ WebCrawlerX ๐
A flexible and efficient web crawler written in Rust.
- Multiple spider implementations (CVE Details, GitHub, Quotes)
- Configurable crawling parameters (delay, concurrent requests, page limit)
- Easy to extend with new spiders
cargo install webcrawlerxList available spiders:
webcrawlerx spidersRun a specific spider:
webcrawlerx run --spider <spider_name>
--spider <spider_name> [--delay <ms>] [--concurrent <num>] [--limit <num>]Example:
webcrawlerx run --spider cvedetails --delay 200 --concurrent 2 --limit 10To add a new spider, create a new module in the spiders directory and implement the Spider trait. Then, update the run_spider function in main.rs to include your new spider.
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.