A spider to crawl links from webpages recursivley with a dir busting option. This tool takes a list of starting URLs, crawls them, finds local links recursively and logs them to Crawled Links.log. Also, it has a DirBusting option.
python3 spider.py -u http://localhost/python3 spider.py --urls <urls_list>
If you want to use the dir busting option:
- Give the code a directories wordlist to work with like so:
python3 spider.py -u http://localhost/ -b directory-list.txt
Setting recusrsion level (how deep the sipder will crawl links)
-
By default, recurse level is set to 5 (5 links to crawl from every page) To change that, all you have to do is use the flag
-ror--recursefollowed by anint. -
Example:
python3 -u http://localhost/ -r 42
- Python3
- BeautifulSoup
The skeleton for the class CrawlerClass | Frankie @ https://github.com/f-prime