Key Features . Installation . Modes/Usage Guide . Next Features . Support . License
httpr is an advanced fully customizable search engines’ OSINT tool designed especially for security researchers and bug-bounty hunters. This tool aims to be simple and scrape search engines using the lowest resources to get the most results possible in the fastest way.
httpr is a tool used for:
- Algorithmic subdomain enumuration
- special keywords enumeration
- web paths enumeration
- Custom dork command enumeration
- 4 Available search engines: Google, Bing, Yahoo & Yandex
- 4 Available modes
- Intelligent Dorking
- Multi-threading by default
- Customizable result output
- Verification support (instead of using httpx)
- IP ban escape
go install -v github.com/thisisiliya/httpr@latestYou can compile httpr from source code.
After unzipping it, use:
go get && go buildYou can install it with:
go installNote that you need at least go 1.18v to compile httpr
docker pull ghcr.io/thisisiliya/httpr-
Or download the Windows version from release
- sub - algorithmic subdomain enumeration for domains
- key - keywords enumeration for domains
- path - path enumeration for domains
- custom - custom dork command to scrape
>>> httpr help sub
algorithmic subdomain enumeration for domain(s)
Usage:
httpr sub [flags]
Examples:
httpr sub --domain hackerone.com --all
Flags:
-a, --all redo the process for the result
--depth int number of pages to scrape per result (default 5)
-d, --domain string target domain to search subdomains
--domains string target domains file path
-h, --help help for sub
--show-sub show subdomains as result
--show-url show URLs as result
>>> httpr sub -d hackerone.com -a --silent
www.hackerone.com
docs.hackerone.com
api.hackerone.com
...
>>> httpr help key
keyword(s) enumeration for domain(s)
Usage:
httpr key [flags]
Examples:
httpr key --domain hackerone.com --keyword report --depth 3
Flags:
--depth int number of pages to scrape per key (default 3)
-d, --domain string target domain to search keyword(s)
--domains string target domains file path
-h, --help help for key
-k, --keyword string target keyword to search
--keywords string target keywords path
--show-host show hosts as result
--show-path show paths as result
--show-sub show subdomains as result
>>> httpr key -d hackerone.com -k report --silent
https://hackerone.com/reports/647130
https://hackerone.com/directory/programs
https://www.hackerone.com/hacker-powered-security-report-0
...
>>> httpr help path
path enumeration for domain(s)
Usage:
httpr path [flags]
Examples:
usage: --domain hackerone.com --depth 10
Flags:
--depth int number of pages to scrape per domain (default 10)
-d, --domain string target domain to search
--domains string target domains file path
-h, --help help for path
--show-path show paths as result
>>> httpr path -d hackerone.com --silent
https://hackerone.com/enter
https://hackerone.com/telegram
https://hackerone.com/rockstargames
...
>>> httpr help custom
engine page(s) scrape by custom dork commands
Usage:
httpr custom [flags]
Examples:
httpr custom --command "site:*.hackerone.com inurl:report" --target-host hackerone.com --engine Google --depth 1
Flags:
-c, --command string dork command to scrape
--depth int number of pages to scrape (default 1)
-e, --engine string target engine to scrape. available engines: Google, Bing, Yahoo (default "Google")
-h, --help help for custom
--show-host show hosts as result
--show-path show paths as result
--show-sub show subdomains as result
--split-by string dork commands split character (default " ")
-t, --target-host string filter result by host
>>> httpr custom -c "site:*.hackerone.com inurl:report" --target-host hackerone.com --silent
https://docs.hackerone.com/en/articles/8475030-report-states
https://docs.hackerone.com/en/articles/8474574-report-actions
https://www.hackerone.com/reports/7th-annual-hacker-powered-security-report
...
- More search engines support
- New modes
If you have any problem with httpr, maybe you can find out your answer on wiki
For any questions, bugs, or assistance, feel free to create an issue
You can support this project with a ⭐
Please visit License file