mirror of
				https://github.com/sqlmapproject/sqlmap.git
				synced 2025-11-04 18:07:46 +03:00 
			
		
		
		
	First patch for #4530
This commit is contained in:
		
							parent
							
								
									f84ec1072b
								
							
						
					
					
						commit
						ccbc3e62f8
					
				| 
						 | 
					@ -18,7 +18,7 @@ from lib.core.enums import OS
 | 
				
			||||||
from thirdparty.six import unichr as _unichr
 | 
					from thirdparty.six import unichr as _unichr
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# sqlmap version (<major>.<minor>.<month>.<monthly commit>)
 | 
					# sqlmap version (<major>.<minor>.<month>.<monthly commit>)
 | 
				
			||||||
VERSION = "1.5.1.12"
 | 
					VERSION = "1.5.1.13"
 | 
				
			||||||
TYPE = "dev" if VERSION.count('.') > 2 and VERSION.split('.')[-1] != '0' else "stable"
 | 
					TYPE = "dev" if VERSION.count('.') > 2 and VERSION.split('.')[-1] != '0' else "stable"
 | 
				
			||||||
TYPE_COLORS = {"dev": 33, "stable": 90, "pip": 34}
 | 
					TYPE_COLORS = {"dev": 33, "stable": 90, "pip": 34}
 | 
				
			||||||
VERSION_STRING = "sqlmap/%s#%s" % ('.'.join(VERSION.split('.')[:-1]) if VERSION.count('.') > 2 and VERSION.split('.')[-1] == '0' else VERSION, TYPE)
 | 
					VERSION_STRING = "sqlmap/%s#%s" % ('.'.join(VERSION.split('.')[:-1]) if VERSION.count('.') > 2 and VERSION.split('.')[-1] == '0' else VERSION, TYPE)
 | 
				
			||||||
| 
						 | 
					@ -757,6 +757,8 @@ MAX_CONNECTION_TOTAL_SIZE = 100 * 1024 * 1024
 | 
				
			||||||
# For preventing MemoryError exceptions (caused when using large sequences in difflib.SequenceMatcher)
 | 
					# For preventing MemoryError exceptions (caused when using large sequences in difflib.SequenceMatcher)
 | 
				
			||||||
MAX_DIFFLIB_SEQUENCE_LENGTH = 10 * 1024 * 1024
 | 
					MAX_DIFFLIB_SEQUENCE_LENGTH = 10 * 1024 * 1024
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					MAX_CHAR_HEURISTICS_SIZE = 10000
 | 
				
			||||||
 | 
					
 | 
				
			||||||
# Maximum (multi-threaded) length of entry in bisection algorithm
 | 
					# Maximum (multi-threaded) length of entry in bisection algorithm
 | 
				
			||||||
MAX_BISECTION_LENGTH = 50 * 1024 * 1024
 | 
					MAX_BISECTION_LENGTH = 50 * 1024 * 1024
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -44,6 +44,7 @@ from lib.core.settings import BLOCKED_IP_REGEX
 | 
				
			||||||
from lib.core.settings import DEFAULT_COOKIE_DELIMITER
 | 
					from lib.core.settings import DEFAULT_COOKIE_DELIMITER
 | 
				
			||||||
from lib.core.settings import EVENTVALIDATION_REGEX
 | 
					from lib.core.settings import EVENTVALIDATION_REGEX
 | 
				
			||||||
from lib.core.settings import IDENTYWAF_PARSE_LIMIT
 | 
					from lib.core.settings import IDENTYWAF_PARSE_LIMIT
 | 
				
			||||||
 | 
					from lib.core.settings import MAX_CHAR_HEURISTICS_SIZE
 | 
				
			||||||
from lib.core.settings import MAX_CONNECTION_TOTAL_SIZE
 | 
					from lib.core.settings import MAX_CONNECTION_TOTAL_SIZE
 | 
				
			||||||
from lib.core.settings import META_CHARSET_REGEX
 | 
					from lib.core.settings import META_CHARSET_REGEX
 | 
				
			||||||
from lib.core.settings import PARSE_HEADERS_LIMIT
 | 
					from lib.core.settings import PARSE_HEADERS_LIMIT
 | 
				
			||||||
| 
						 | 
					@ -258,7 +259,7 @@ def getHeuristicCharEncoding(page):
 | 
				
			||||||
    """
 | 
					    """
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    key = hash(page)
 | 
					    key = hash(page)
 | 
				
			||||||
    retVal = kb.cache.encoding.get(key) or detect(page)["encoding"]
 | 
					    retVal = kb.cache.encoding.get(key) or detect(page[:MAX_CHAR_HEURISTICS_SIZE])["encoding"]
 | 
				
			||||||
    kb.cache.encoding[key] = retVal
 | 
					    kb.cache.encoding[key] = retVal
 | 
				
			||||||
 | 
					
 | 
				
			||||||
    if retVal and retVal.lower().replace('-', "") == UNICODE_ENCODING.lower().replace('-', ""):
 | 
					    if retVal and retVal.lower().replace('-', "") == UNICODE_ENCODING.lower().replace('-', ""):
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in New Issue
	
	Block a user