Skip to content

๐Ÿ•ท๏ธ WebCrawlerX ๐Ÿš€ is a rust based crawler for the open web ,inspired by scrapy.

License

Notifications You must be signed in to change notification settings

mahmudsudo/-WebCrawlerX-

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

4 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

๐Ÿ•ท๏ธ WebCrawlerX ๐Ÿš€

A flexible and efficient web crawler written in Rust.

Features

  • Multiple spider implementations (CVE Details, GitHub, Quotes)
  • Configurable crawling parameters (delay, concurrent requests, page limit)
  • Easy to extend with new spiders

Installation

cargo install webcrawlerx

Usage

List available spiders:

webcrawlerx spiders

Run a specific spider:

webcrawlerx run --spider <spider_name>
--spider <spider_name> [--delay <ms>] [--concurrent <num>] [--limit <num>]

Example:

webcrawlerx run --spider cvedetails --delay 200 --concurrent 2 --limit 10

Adding a New Spider

To add a new spider, create a new module in the spiders directory and implement the Spider trait. Then, update the run_spider function in main.rs to include your new spider.

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

License

This project is licensed under the MIT License - see the LICENSE file for details.

About

๐Ÿ•ท๏ธ WebCrawlerX ๐Ÿš€ is a rust based crawler for the open web ,inspired by scrapy.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages