forked from faizann24/XssPy
-
Notifications
You must be signed in to change notification settings - Fork 0
/
XssPy.py
197 lines (175 loc) · 7.67 KB
/
XssPy.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
#!/usr/bin/env python
import mechanize
import sys
import httplib
import argparse
import logging
from urlparse import urlparse
br = mechanize.Browser() # initiating the browser
br.addheaders = [
('User-agent',
'Mozilla/5.0 (Windows; U; Windows NT 5.1; it; rv:1.8.1.11)Gecko/20071127 Firefox/2.0.0.11')
]
br.set_handle_robots(False)
br.set_handle_refresh(False)
payloads = ['<svg "ons>', '" onfocus="alert(1);', 'javascript:alert(1)']
blacklist = ['.png', '.jpg', '.jpeg', '.mp3', '.mp4', '.avi', '.gif', '.svg',
'.pdf']
xssLinks = [] # TOTAL CROSS SITE SCRIPTING FINDINGS
class color:
BLUE = '\033[94m'
RED = '\033[91m'
GREEN = '\033[92m'
YELLOW = '\033[93m'
BOLD = '\033[1m'
END = '\033[0m'
@staticmethod
def log(lvl, col, msg):
logger.log(lvl, col + msg + color.END)
print color.BOLD + color.RED + """
XssPy - Finding XSS made easier
Author: Faizan Ahmad (Fsecurify)
Email: fsecurify@gmail.com
Usage: XssPy.py website.com (Not www.website.com OR http://www.website.com)
Comprehensive Scan: python XssPy.py -u website.com -e
Verbose logging: python XssPy.py -u website.com -v
Cookies: python XssPy.py -u website.complex -c name=val name=val
Description: XssPy is a python tool for finding Cross Site Scripting
vulnerabilities in websites. This tool is the first of its kind.
Instead of just checking one page as most of the tools do, this tool
traverses the website and find all the links and subdomains first.
After that, it starts scanning each and every input on each and every
page that it found while its traversal. It uses small yet effective
payloads to search for XSS vulnerabilities. XSS in many high
profile websites and educational institutes has been found
by using this very tool.
""" + color.END
logger = logging.getLogger(__name__)
lh = logging.StreamHandler() # Handler for the logger
logger.addHandler(lh)
formatter = logging.Formatter('[%(asctime)s] %(message)s', datefmt='%H:%M:%S')
lh.setFormatter(formatter)
parser = argparse.ArgumentParser()
parser.add_argument('-u', action='store', dest='url',
help='The URL to analyze')
parser.add_argument('-e', action='store_true', dest='compOn',
help='Enable comprehensive scan')
parser.add_argument('-v', action='store_true', dest='verbose',
help='Enable verbose logging')
parser.add_argument('-c', action='store', dest='cookies',
help='Space separated list of cookies',
nargs='+', default=[])
results = parser.parse_args()
logger.setLevel(logging.DEBUG if results.verbose else logging.INFO)
def testPayload(payload, p, link):
br.form[str(p.name)] = payload
br.submit()
# if payload is found in response, we have XSS
if payload in br.response().read():
color.log(logging.DEBUG, color.BOLD + color.GREEN, 'XSS found!')
report = 'Link: %s, Payload: %s, Element: %s' % (str(link),
payload, str(p.name))
color.log(logging.INFO, color.BOLD + color.GREEN, report)
xssLinks.append(report)
br.back()
def initializeAndFind():
if not results.url: # if the url has been passed or not
color.log(logging.INFO, color.GREEN, 'Url not provided correctly')
return []
firstDomains = [] # list of domains
allURLS = []
allURLS.append(results.url) # just one url at the moment
largeNumberOfUrls = [] # in case one wants to do comprehensive search
# doing a short traversal if no command line argument is being passed
color.log(logging.INFO, color.GREEN, 'Doing a short traversal.')
for url in allURLS:
smallurl = str(url)
# Test HTTPS/HTTP compatibility. Prefers HTTPS but defaults to
# HTTP if any errors are encountered
try:
test = httplib.HTTPSConnection(smallurl)
test.request("GET", "/")
response = test.getresponse()
if (response.status == 200) | (response.status == 302):
url = "https://www." + str(url)
elif response.status == 301:
loc = response.getheader('Location')
url = loc.scheme + '://' + loc.netloc
else:
url = "http://www." + str(url)
except:
url = "http://www." + str(url)
try:
br.open(url)
for cookie in results.cookies:
color.log(logging.INFO, color.BLUE,
'Adding cookie: %s' % cookie)
br.set_cookie(cookie)
br.open(url)
color.log(logging.INFO, color.GREEN,
'Finding all the links of the website ' + str(url))
for link in br.links(): # finding the links of the website
if smallurl in str(link.absolute_url):
firstDomains.append(str(link.absolute_url))
firstDomains = list(set(firstDomains))
except:
pass
color.log(logging.INFO, color.GREEN,
'Number of links to test are: ' + str(len(firstDomains)))
if results.compOn:
color.log(logging.INFO, color.GREEN,
'Doing a comprehensive traversal. This may take a while')
for link in firstDomains:
try:
br.open(link)
# going deeper into each link and finding its links
for newlink in br.links():
if smallurl in str(newlink.absolute_url):
largeNumberOfUrls.append(newlink.absolute_url)
except:
pass
firstDomains = list(set(firstDomains + largeNumberOfUrls))
color.log(logging.INFO, color.GREEN,
'Total Number of links to test have become: ' +
str(len(firstDomains))) # all links have been found
return firstDomains
def findxss(firstDomains):
# starting finding XSS
color.log(logging.INFO, color.GREEN, 'Started finding XSS')
if firstDomains: # if there is atleast one link
for link in firstDomains:
blacklisted = False
y = str(link)
color.log(logging.DEBUG, color.YELLOW, str(link))
for ext in blacklist:
if ext in y:
color.log(logging.DEBUG, color.RED,
'\tNot a good url to test')
blacklisted = True
break
if not blacklisted:
try:
br.open(str(link)) # open the link
if br.forms(): # if a form exists, submit it
params = list(br.forms())[0] # our form
br.select_form(nr=0) # submit the first form
for p in params.controls:
par = str(p)
# submit only those forms which require text
if 'TextControl' in par:
color.log(logging.DEBUG, color.YELLOW,
'\tParam: ' + str(p.name))
for item in payloads:
testPayload(item, p, link)
except:
pass
color.log(logging.DEBUG, color.GREEN + color.BOLD,
'The following links are vulnerable: ')
for link in xssLinks: # print all xss findings
color.log(logging.DEBUG, color.GREEN, '\t' + link)
else:
color.log(logging.INFO, color.RED + color.BOLD,
'\tNo link found, exiting')
# calling the function
firstDomains = initializeAndFind()
findxss(firstDomains)