sqlmap/lib/utils/google.py

184 lines
7.2 KiB
Python
Raw Normal View History

#!/usr/bin/env python
2008-10-15 19:38:22 +04:00
"""
2015-01-06 17:02:16 +03:00
Copyright (c) 2006-2015 sqlmap developers (http://sqlmap.org/)
2010-10-15 03:18:29 +04:00
See the file 'doc/COPYING' for copying permission
2008-10-15 19:38:22 +04:00
"""
import cookielib
2010-10-31 18:41:28 +03:00
import httplib
2008-10-15 19:38:22 +04:00
import re
import socket
2012-07-05 14:34:27 +04:00
import urllib
2008-10-15 19:38:22 +04:00
import urllib2
from lib.core.common import getSafeExString
2010-06-02 16:45:40 +04:00
from lib.core.common import getUnicode
from lib.core.common import readInput
2012-07-31 13:03:44 +04:00
from lib.core.common import urlencode
2008-10-15 19:38:22 +04:00
from lib.core.data import conf
from lib.core.data import logger
2012-12-07 14:54:34 +04:00
from lib.core.enums import CUSTOM_LOGGING
from lib.core.enums import HTTP_HEADER
from lib.core.exception import SqlmapConnectionException
from lib.core.exception import SqlmapGenericException
2015-08-31 11:24:05 +03:00
from lib.core.exception import SqlmapUserQuitException
from lib.core.settings import DUMMY_SEARCH_USER_AGENT
from lib.core.settings import DUCKDUCKGO_REGEX
2015-08-31 11:24:05 +03:00
from lib.core.settings import DISCONNECT_SEARCH_REGEX
from lib.core.settings import GOOGLE_REGEX
from lib.core.settings import HTTP_ACCEPT_ENCODING_HEADER_VALUE
2011-01-30 14:36:03 +03:00
from lib.core.settings import UNICODE_ENCODING
from lib.request.basic import decodePage
from lib.request.httpshandler import HTTPSHandler
2015-09-27 17:09:02 +03:00
from thirdparty.socks import socks
2008-10-15 19:38:22 +04:00
class Google(object):
2008-10-15 19:38:22 +04:00
"""
This class defines methods used to perform Google dorking (command
2015-08-14 11:18:51 +03:00
line option '-g <google dork>')
2008-10-15 19:38:22 +04:00
"""
def __init__(self, handlers):
2012-03-07 18:48:45 +04:00
self._cj = cookielib.CookieJar()
2012-02-20 14:02:19 +04:00
handlers.append(urllib2.HTTPCookieProcessor(self._cj))
handlers.append(HTTPSHandler())
self.opener = urllib2.build_opener(*handlers)
2008-10-15 19:38:22 +04:00
self.opener.addheaders = conf.httpHeaders
try:
conn = self.opener.open("https://www.google.com/ncr")
conn.info() # retrieve session cookie
2014-10-28 17:34:53 +03:00
except Exception, ex:
errMsg = "unable to connect to Google ('%s')" % getSafeExString(ex)
raise SqlmapConnectionException(errMsg)
2008-10-15 19:38:22 +04:00
def search(self, dork):
2008-10-15 19:38:22 +04:00
"""
This method performs the effective search on Google providing
the google dork and the Google session cookie
"""
gpage = conf.googlePage if conf.googlePage > 1 else 1
2010-03-05 17:59:33 +03:00
logger.info("using Google result page #%d" % gpage)
2010-09-27 17:41:18 +04:00
if not dork:
2008-10-15 19:38:22 +04:00
return None
url = "https://www.google.com/search?"
url += "q=%s&" % urlencode(dork, convall=True)
2011-11-06 15:42:02 +04:00
url += "num=100&hl=en&complete=0&safe=off&filter=0&btnG=Search"
2013-01-10 16:18:44 +04:00
url += "&start=%d" % ((gpage - 1) * 100)
2008-10-15 19:38:22 +04:00
try:
conn = self.opener.open(url)
2010-10-31 18:41:28 +03:00
requestMsg = "HTTP request:\nGET %s" % url
requestMsg += " %s" % httplib.HTTPConnection._http_vsn_str
2012-12-07 14:54:34 +04:00
logger.log(CUSTOM_LOGGING.TRAFFIC_OUT, requestMsg)
page = conn.read()
code = conn.code
status = conn.msg
responseHeaders = conn.info()
page = decodePage(page, responseHeaders.get("Content-Encoding"), responseHeaders.get("Content-Type"))
responseMsg = "HTTP response (%s - %d):\n" % (status, code)
2010-09-27 17:41:18 +04:00
if conf.verbose <= 4:
2011-01-30 14:36:03 +03:00
responseMsg += getUnicode(responseHeaders, UNICODE_ENCODING)
elif conf.verbose > 4:
responseMsg += "%s\n%s\n" % (responseHeaders, page)
2010-09-27 17:41:18 +04:00
2012-12-07 14:54:34 +04:00
logger.log(CUSTOM_LOGGING.TRAFFIC_IN, responseMsg)
2008-10-15 19:38:22 +04:00
except urllib2.HTTPError, e:
2010-05-16 00:44:08 +04:00
try:
page = e.read()
except socket.timeout:
2011-04-30 17:20:05 +04:00
warnMsg = "connection timed out while trying "
2010-05-16 00:44:08 +04:00
warnMsg += "to get error page information (%d)" % e.code
2010-09-27 17:41:18 +04:00
logger.critical(warnMsg)
2010-05-16 00:44:08 +04:00
return None
2015-09-27 17:09:02 +03:00
except (urllib2.URLError, httplib.error, socket.error, socket.timeout, socks.ProxyError):
2008-10-15 19:38:22 +04:00
errMsg = "unable to connect to Google"
raise SqlmapConnectionException(errMsg)
2008-10-15 19:38:22 +04:00
retVal = [urllib.unquote(match.group(1)) for match in re.finditer(GOOGLE_REGEX, page, re.I | re.S)]
2011-02-08 03:02:54 +03:00
if not retVal and "detected unusual traffic" in page:
2011-04-30 17:20:05 +04:00
warnMsg = "Google has detected 'unusual' traffic from "
2013-03-18 14:41:15 +04:00
warnMsg += "used IP address disabling further searches"
raise SqlmapGenericException(warnMsg)
2008-10-15 19:38:22 +04:00
if not retVal:
2015-08-31 11:24:05 +03:00
message = "no usable links found. What do you want to do?"
message += "\n[1] (re)try with DuckDuckGo (default)"
message += "\n[2] (re)try with Disconnect Search"
message += "\n[3] quit"
choice = readInput(message, default="1").strip().upper()
if choice == "Q":
raise SqlmapUserQuitException
elif choice == "2":
url = "https://search.disconnect.me/searchTerms/search?"
url += "start=nav&option=Web"
url += "&query=%s" % urlencode(dork, convall=True)
url += "&ses=Google&location_option=US"
url += "&nextDDG=%s" % urlencode("/search?q=&num=100&hl=en&start=%d&sa=N" % ((gpage - 1) * 10), convall=True)
url += "&sa=N&showIcons=false&filterIcons=none&js_enabled=1"
regex = DISCONNECT_SEARCH_REGEX
else:
url = "https://duckduckgo.com/d.js?"
url += "q=%s&p=%d&s=100" % (urlencode(dork, convall=True), gpage)
2015-08-31 11:24:05 +03:00
regex = DUCKDUCKGO_REGEX
if not conf.randomAgent:
self.opener.addheaders = [_ for _ in self.opener.addheaders if _[0].lower() != HTTP_HEADER.USER_AGENT.lower()]
self.opener.addheaders.append((HTTP_HEADER.USER_AGENT, DUMMY_SEARCH_USER_AGENT))
self.opener.addheaders = [_ for _ in self.opener.addheaders if _[0].lower() != HTTP_HEADER.ACCEPT_ENCODING.lower()]
self.opener.addheaders.append((HTTP_HEADER.ACCEPT_ENCODING, HTTP_ACCEPT_ENCODING_HEADER_VALUE))
try:
conn = self.opener.open(url)
requestMsg = "HTTP request:\nGET %s" % url
requestMsg += " %s" % httplib.HTTPConnection._http_vsn_str
logger.log(CUSTOM_LOGGING.TRAFFIC_OUT, requestMsg)
page = conn.read()
code = conn.code
status = conn.msg
responseHeaders = conn.info()
page = decodePage(page, responseHeaders.get("Content-Encoding"), responseHeaders.get("Content-Type"))
2015-08-31 11:24:05 +03:00
responseMsg = "HTTP response (%s - %d):\n" % (status, code)
2015-08-31 11:24:05 +03:00
if conf.verbose <= 4:
responseMsg += getUnicode(responseHeaders, UNICODE_ENCODING)
elif conf.verbose > 4:
responseMsg += "%s\n%s\n" % (responseHeaders, page)
2015-08-31 11:24:05 +03:00
logger.log(CUSTOM_LOGGING.TRAFFIC_IN, responseMsg)
except urllib2.HTTPError, e:
try:
2015-08-31 11:24:05 +03:00
page = e.read()
except socket.timeout:
warnMsg = "connection timed out while trying "
warnMsg += "to get error page information (%d)" % e.code
logger.critical(warnMsg)
return None
except:
errMsg = "unable to connect"
raise SqlmapConnectionException(errMsg)
retVal = [urllib.unquote(match.group(1)) for match in re.finditer(regex, page, re.I | re.S)]
return retVal
def setHTTPProxy(): # Cross-linked function
raise NotImplementedError