mirror of
				https://github.com/sqlmapproject/sqlmap.git
				synced 2025-10-31 16:07:55 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			229 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			229 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| #!/usr/bin/env python
 | |
| 
 | |
| """
 | |
| $Id$
 | |
| 
 | |
| Copyright (c) 2006-2012 sqlmap developers (http://www.sqlmap.org/)
 | |
| See the file 'doc/COPYING' for copying permission
 | |
| """
 | |
| 
 | |
| import codecs
 | |
| import gzip
 | |
| import logging
 | |
| import re
 | |
| import StringIO
 | |
| import zlib
 | |
| 
 | |
| from extra.chardet import detect
 | |
| from lib.core.common import extractErrorMessage
 | |
| from lib.core.common import extractRegexResult
 | |
| from lib.core.common import getUnicode
 | |
| from lib.core.common import readInput
 | |
| from lib.core.common import resetCookieJar
 | |
| from lib.core.common import singleTimeLogMessage
 | |
| from lib.core.data import conf
 | |
| from lib.core.data import kb
 | |
| from lib.core.data import logger
 | |
| from lib.core.enums import HTTPHEADER
 | |
| from lib.core.enums import PLACE
 | |
| from lib.core.settings import DEFAULT_COOKIE_DELIMITER
 | |
| from lib.core.settings import ML
 | |
| from lib.core.settings import META_CHARSET_REGEX
 | |
| from lib.core.settings import PARSE_HEADERS_LIMIT
 | |
| from lib.parse.headers import headersParser
 | |
| from lib.parse.html import htmlParser
 | |
| 
 | |
| def forgeHeaders(items=None):
 | |
|     """
 | |
|     Prepare HTTP Cookie, HTTP User-Agent and HTTP Referer headers to use when performing
 | |
|     the HTTP requests
 | |
|     """
 | |
| 
 | |
|     items = items or {}
 | |
| 
 | |
|     for _ in items.keys():
 | |
|         if items[_] is None:
 | |
|             del items[_]
 | |
| 
 | |
|     headers = dict(conf.httpHeaders)
 | |
|     headers.update(items or {})
 | |
| 
 | |
|     if conf.cj:
 | |
|         if HTTPHEADER.COOKIE in headers:
 | |
|             for cookie in conf.cj:
 | |
|                 if ("%s=" % cookie.name) in headers[HTTPHEADER.COOKIE]:
 | |
|                     if kb.mergeCookies is None:
 | |
|                         message = "you provided a HTTP %s header value. " % HTTPHEADER.COOKIE
 | |
|                         message += "The target url provided its own cookies within "
 | |
|                         message += "the HTTP %s header which intersect with yours. " % HTTPHEADER.SET_COOKIE
 | |
|                         message += "Do you want to merge them in futher requests? [Y/n] "
 | |
|                         test = readInput(message, default="Y")
 | |
|                         kb.mergeCookies = not test or test[0] in ("y", "Y")
 | |
| 
 | |
|                     if kb.mergeCookies:
 | |
|                         _ = lambda x: re.sub("(?i)%s=[^%s]+" % (cookie.name, DEFAULT_COOKIE_DELIMITER), "%s=%s" % (cookie.name, cookie.value), x)
 | |
|                         headers[HTTPHEADER.COOKIE] = _(headers[HTTPHEADER.COOKIE])
 | |
| 
 | |
|                         if PLACE.COOKIE in conf.parameters:
 | |
|                             conf.parameters[PLACE.COOKIE] = _(conf.parameters[PLACE.COOKIE])
 | |
|                         conf.httpHeaders = [(item[0], item[1] if item[0] != HTTPHEADER.COOKIE else _(item[1])) for item in conf.httpHeaders]
 | |
| 
 | |
|                 elif not kb.testMode:
 | |
|                     headers[HTTPHEADER.COOKIE] += "%s %s=%s" % (DEFAULT_COOKIE_DELIMITER, cookie.name, cookie.value)
 | |
| 
 | |
|         if kb.testMode:
 | |
|             resetCookieJar(conf.cj)
 | |
| 
 | |
|     if kb.redirectSetCookie and not conf.dropSetCookie:
 | |
|         if HTTPHEADER.COOKIE in headers:
 | |
|             headers[HTTPHEADER.COOKIE] += "%s %s" % (DEFAULT_COOKIE_DELIMITER, kb.redirectSetCookie)
 | |
|         else:
 | |
|             headers[HTTPHEADER.COOKIE] = kb.redirectSetCookie
 | |
| 
 | |
|     return headers
 | |
| 
 | |
| def parseResponse(page, headers):
 | |
|     """
 | |
|     @param page: the page to parse to feed the knowledge base htmlFp
 | |
|     (back-end DBMS fingerprint based upon DBMS error messages return
 | |
|     through the web application) list and absFilePaths (absolute file
 | |
|     paths) set.
 | |
|     """
 | |
| 
 | |
|     if headers:
 | |
|         headersParser(headers)
 | |
| 
 | |
|     if page:
 | |
|         htmlParser(page)
 | |
| 
 | |
| def checkCharEncoding(encoding):
 | |
|     if encoding:
 | |
|         encoding = encoding.lower()
 | |
|     else:
 | |
|         return encoding
 | |
| 
 | |
|     # http://www.destructor.de/charsets/index.htm
 | |
|     translate = { "windows-874": "iso-8859-11", "en_us": "utf8", "macintosh": "iso-8859-1", "euc_tw": "big5_tw", "th": "tis-620", "unicode": "utf8" }
 | |
| 
 | |
|     for delimiter in (';', ',', '('):
 | |
|         if delimiter in encoding:
 | |
|             encoding = encoding[:encoding.find(delimiter)].strip()
 | |
| 
 | |
|     # popular typos/errors
 | |
|     if "8858" in encoding:
 | |
|         encoding = encoding.replace("8858", "8859") # iso-8858 -> iso-8859
 | |
|     elif "8559" in encoding:
 | |
|         encoding = encoding.replace("8559", "8859") # iso-8559 -> iso-8859
 | |
|     elif "5889" in encoding:
 | |
|         encoding = encoding.replace("5889", "8859") # iso-5889 -> iso-8859
 | |
|     elif "2313" in encoding:
 | |
|         encoding = encoding.replace("2313", "2312") # gb2313 -> gb2312
 | |
|     elif "x-euc" in encoding:
 | |
|         encoding = encoding.replace("x-euc", "euc") # x-euc-kr -> euc-kr
 | |
| 
 | |
|     # name adjustment for compatibility
 | |
|     if encoding.startswith("8859"):
 | |
|         encoding = "iso-%s" % encoding
 | |
|     elif encoding.startswith("cp-"):
 | |
|         encoding = "cp%s" % encoding[3:]
 | |
|     elif encoding.startswith("euc-"):
 | |
|         encoding = "euc_%s" % encoding[4:]
 | |
|     elif encoding.startswith("windows") and not encoding.startswith("windows-"):
 | |
|         encoding = "windows-%s" % encoding[7:]
 | |
|     elif encoding.find("iso-88") > 0:
 | |
|         encoding = encoding[encoding.find("iso-88"):]
 | |
|     elif encoding.startswith("is0-"):
 | |
|         encoding = "iso%s" % encoding[4:]
 | |
|     elif encoding.find("ascii") > 0:
 | |
|         encoding = "ascii"
 | |
|     elif encoding.find("utf8") > 0:
 | |
|         encoding = "utf8"
 | |
| 
 | |
|     # http://philip.html5.org/data/charsets-2.html
 | |
|     if encoding in translate:
 | |
|         encoding = translate[encoding]
 | |
|     elif encoding in ("null", "{charset}", "*"):
 | |
|         return None
 | |
| 
 | |
|     # http://www.iana.org/assignments/character-sets
 | |
|     # http://docs.python.org/library/codecs.html
 | |
|     try:
 | |
|         codecs.lookup(encoding)
 | |
|     except LookupError:
 | |
|         warnMsg = "unknown web page charset '%s'. " % encoding
 | |
|         warnMsg += "Please report by e-mail to %s." % ML
 | |
|         singleTimeLogMessage(warnMsg, logging.WARN, encoding)
 | |
|         encoding = None
 | |
| 
 | |
|     return encoding
 | |
| 
 | |
| def getHeuristicCharEncoding(page):
 | |
|     """
 | |
|     Returns page encoding charset detected by usage of heuristics
 | |
|     Reference: http://chardet.feedparser.org/docs/
 | |
|     """
 | |
|     retVal = detect(page)["encoding"]
 | |
| 
 | |
|     infoMsg = "heuristics detected web page charset '%s'" % retVal
 | |
|     singleTimeLogMessage(infoMsg, logging.INFO, retVal)
 | |
| 
 | |
|     return retVal
 | |
| 
 | |
| def decodePage(page, contentEncoding, contentType):
 | |
|     """
 | |
|     Decode compressed/charset HTTP response
 | |
|     """
 | |
| 
 | |
|     if not page or (conf.nullConnection and len(page) < 2):
 | |
|         return getUnicode(page)
 | |
| 
 | |
|     if isinstance(contentEncoding, basestring) and contentEncoding.lower() in ("gzip", "x-gzip", "deflate"):
 | |
|         if contentEncoding == "deflate":
 | |
|             # http://stackoverflow.com/questions/1089662/python-inflate-and-deflate-implementations
 | |
|             data = StringIO.StringIO(zlib.decompress(page, -15))
 | |
|         else:
 | |
|             data = gzip.GzipFile('', 'rb', 9, StringIO.StringIO(page))
 | |
| 
 | |
|         try:
 | |
|             page = data.read()
 | |
|         except Exception, msg:
 | |
|             errMsg = "detected invalid data for declared content "
 | |
|             errMsg += "encoding '%s' ('%s')" % (contentEncoding, msg)
 | |
|             singleTimeLogMessage(errMsg, logging.ERROR)
 | |
| 
 | |
|     if not conf.charset:
 | |
|         httpCharset, metaCharset = None, None
 | |
| 
 | |
|         # http://stackoverflow.com/questions/1020892/python-urllib2-read-to-unicode
 | |
|         if contentType and (contentType.find('charset=') != -1):
 | |
|             httpCharset = checkCharEncoding(contentType.split('charset=')[-1])
 | |
| 
 | |
|         metaCharset = checkCharEncoding(extractRegexResult(META_CHARSET_REGEX, page, re.DOTALL | re.IGNORECASE))
 | |
| 
 | |
|         if ((httpCharset or metaCharset) and not all([httpCharset, metaCharset]))\
 | |
|             or (httpCharset == metaCharset and all([httpCharset, metaCharset])):
 | |
|             kb.pageEncoding = httpCharset or metaCharset
 | |
|         else:
 | |
|             kb.pageEncoding = None
 | |
|     else:
 | |
|         kb.pageEncoding = conf.charset
 | |
| 
 | |
|     if contentType and not isinstance(page, unicode) and any(map(lambda x: x in contentType.lower(), ('text/txt', 'text/raw', 'text/html', 'text/xml'))):
 | |
|         # can't do for all responses because we need to support binary files too
 | |
|         kb.pageEncoding = kb.pageEncoding or checkCharEncoding(getHeuristicCharEncoding(page))
 | |
|         page = getUnicode(page, kb.pageEncoding)
 | |
| 
 | |
|     return page
 | |
| 
 | |
| def processResponse(page, responseHeaders):
 | |
|     kb.processResponseCounter += 1
 | |
| 
 | |
|     if not kb.dumpMode:
 | |
|         parseResponse(page, responseHeaders if kb.processResponseCounter < PARSE_HEADERS_LIMIT else None)
 | |
| 
 | |
|     if conf.parseErrors:
 | |
|         msg = extractErrorMessage(page)
 | |
| 
 | |
|         if msg:
 | |
|             logger.info("parsed error message: '%s'" % msg)
 |