Finding URLs
Let’s find URLs from live hosts using waybackurls, gau, and katana.
We will feed the find URLs to automated tools such as gf to test for XSS and stuff.
waybackurls
httprobe_result.txt contains the list of live hosts that is verified through httprobe.
I feed this to waybackurls to find URLs:
cat subs.txt | /home/carabiner1/go/bin/waybackurls| sudo /home/carabiner1/go/bin/anew waybackurls.txt

waybackurls found 13543 URLs:

gau
Now let’s use gau to collect more URLs:
cat subs.txt | /home/carabiner1/go/bin/gau| /home/carabiner1/go/bin/anew gau.txt

gau somehow just froze. Some people say only using katana is also fine.
katana
Based on my experience,
katanais the fastest.
Make sure before doing this Burp Bounty and AutoRepeater is set up for SSRF and Blind XSS!
Now let’s use katana to collect more URLs:
while read subdomain; do /home/carabiner1/go/bin/katana -proxy http://127.0.0.1:8080 -jc -aff -u "$subdomain" | sudo /home/carabiner1/go/bin/anew katana.txt; done < subs.txt
I am also having katana go through the burp proxy so that I can log more information on Burp Suite Targets. This also eases on testing SSRF and XSS automatically.

I stopped it running in the middle of the scan and yet it still discovered 50003 URLs:

Sort It
Let’s sort for unique URLs:
cat *.txt | sort -u | sudo tee all_urls.txt
Now we have 63528 unique URLs to be tested:
