Multithreaded and recursive directory scraping script. Stores useful information with the filepath and finding. All in one file, no external packages required!
python3 hamburglar.py -w -v -h path
Directory Traversal
python3 hamburglar.py ~/Directory/
- This will recursively scan for files in the given directory, then analyzes each file for a variety of findings using regex filters
Single File Analysis
python3 hamburglar.py ~/Directory/file.txt
- This will recursively scan for files in the given directory, then analyzes each file for a variety of findings using regex filters
Web Request Mode
python3 hamburglar.py -w https://google.com
- Adding a
-w
to hamburgler.py tells the script to handle the path as a url. - Currently this does not spider the page, it just analyzes the requested html content
- Adding a
Tips
- Adding
-v
will set the script into verbose mode, and-h
will show details of available arguments - Adding
-o FILENAME
will set the results filename, this is especially useful in scripting situations where you might want multiple results tables (ie github repo spidering)
whitelistOn
: turns on or off whitelist checkingmaxWorkers
: number of worker threads to run concurrently when reading file stackwhitelist
: list of files or directories to exclusively scan for (if whitelistOn=True)blacklist
: list of files, extensions, or directories to block in scanregexList
: dictionary of regex filters with filter type as the key
- ipv4 addresses (public and local)
- emails
- private keys
- urls
- cryptocurrency addresses
{
"/home/adam/Dev/test/email.txt": {
"emails": "{'[email protected]'}"
},
"/home/adam/Dev/test/email2.txt": {
"emails": "{'[email protected]'}"
},
"/home/adam/Dev/test/ips.txt": {
"ipv4": "{'10.0.11.2', '192.168.1.1'}"
},
"/home/adam/Dev/test/test2/email.txt": {
"emails": "{'[email protected]', '[email protected]'}"
},
"/home/adam/Dev/test/test2/ips.txt": {
"ipv4": "{'10.0.11.2', '192.168.1.1'}"
},
"/home/adam/Dev/test/test2/links.txt": {
"site": "{'http://login.web.com'}"
}
}
-
Inspiration came from needmorecowbell/sniff-paste, I wanted the same regex scraping but for every file in a given directory.
-
Please contribute! If there's an error let me know -- even better if you can fix it :)
- Regex Contributions would be very helpful, and should be pretty easy to add!
-
Please don't use this project maliciously, it is meant to be an analysis tool