DE
devanshbatham/ParamSpider
Mining URLs from dark corners of Web Archives for bug hunting/fuzzing/further probing
paramspider
Mining URLs from dark corners of Web Archives for bug hunting/fuzzing/further probing
๐ About โข ๐๏ธ Installation โข โ๏ธ Usage โข ๐ Examples โข ๐ค Contributing โข
About
paramspider allows you to fetch URLs related to any domain or a list of domains from Wayback achives. It filters out "boring" URLs, allowing you to focus on the ones that matter the most.
Installation
To install paramspider, follow these steps:
git clone https://github.com/devanshbatham/paramspider
cd paramspider
pip install .Usage
To use paramspider, follow these steps:
paramspider -d example.comExamples
Here are a few examples of how to use paramspider:
-
Discover URLs for a single domain:
paramspider -d example.com
-
Discover URLs for multiple domains from a file:
paramspider -l domains.txt
-
Stream URLs on the termial:
paramspider -d example.com -s
-
Set up web request proxy:
paramspider -d example.com --proxy '127.0.0.1:7890' -
Adding a placeholder for URL parameter values (default: "FUZZ"):
paramspider -d example.com -p '"><h1>reflection</h1>'
Star History
On this page
Languages
Python95.5%Dockerfile4.5%
MIT License
Created April 12, 2020
Updated March 8, 2026
