a script which can be used for recon and exploitation purpose and is a combination of various tools . this tool works only on kali for oither distributions use docker
Install
git clone https://github.com/harshnandwana/hussh
Docker
$ docker run -it -d --name hussh shubham9672/hussh
Now whenever you want to run your Docker container just use given below command
$ docker exec -it hussh /bin/bash
Commad for using hussh in docker
$ cd hussh
$ ./hussh <target.com>
Usage
$ chmod +x *
$ ./hussh <target.com>
Dependencies along with all files present here use
$ sudo apt-get install sublist3r
$ sudo apt-get install subfinder
$ sudo apt-get install gospider
$ sudo apt-get install golang
Output
To see the output move to /target/hussh-date/ and track for the file target.html and open it with your browser and you will se a page like this.
The process of work is shown here feel free to do changes and commit here.
what this tool do.
- it gathers subdomain using
- sublist3r
- subfinder append these data to a text file namely combined.txt in the folder /target/hussh-date/combined.txt and trim the duplicasy
- check for active subdomains
- check all the active domains in the combined.txt using a tool httprobe by tomnomnom and then keep only https:// sources there as these are the ultimate target.
- use a go script named as go to find out the status code of all the domains this script is also the modified version of a script named _ by tomnomnom and then grep out the url having status code 200 and 302. append these data to a file named a as usefull.txt
- way back machine
- check for the url from usefull.txt in wyback machine and writesd the output in /target/hussh-date/wb.txt
- now using a file inturl which is also inspired from _ by tomnomnom it sorts out only intersting files from web.txt
- spider the subdomains
- using gospider aas my long companion to spider the subdomains and find all other subdomains that may be available there
- now filtering out the usfull data such as sites with url|robots|javascript|subdomains|forms using grep and appending them to a file /target/hussh-date/vulnd.txt
- finding intersting files
- appending usefull.txt and vulnd.txt to a file name vuln.txt and removing the duplicates.
- again finding interesting files from vuln.txt
- hunting or vulnerability
- hunt for cors vulnerability in vulns.txt using cors script by tomnomnom
please add other vulnerability and fork if using