mirror of
https://github.com/sqlmapproject/sqlmap.git
synced 2024-11-25 19:13:48 +03:00
some more refactoring
This commit is contained in:
parent
facce2c0df
commit
d2dd47fb23
|
@ -1645,15 +1645,12 @@ def readCachedFileContent(filename, mode='rb'):
|
|||
"""
|
||||
|
||||
if filename not in kb.cache.content:
|
||||
kb.locks.cache.acquire()
|
||||
|
||||
if filename not in kb.cache.content:
|
||||
checkFile(filename)
|
||||
with codecs.open(filename, mode, UNICODE_ENCODING) as f:
|
||||
content = f.read()
|
||||
kb.cache.content[filename] = content
|
||||
|
||||
kb.locks.cache.release()
|
||||
with kb.locks.cache:
|
||||
if filename not in kb.cache.content:
|
||||
checkFile(filename)
|
||||
with codecs.open(filename, mode, UNICODE_ENCODING) as f:
|
||||
content = f.read()
|
||||
kb.cache.content[filename] = content
|
||||
|
||||
return kb.cache.content[filename]
|
||||
|
||||
|
@ -2113,13 +2110,10 @@ def logHTTPTraffic(requestLogMsg, responseLogMsg):
|
|||
if not conf.trafficFile:
|
||||
return
|
||||
|
||||
kb.locks.log.acquire()
|
||||
|
||||
dataToTrafficFile("%s%s" % (requestLogMsg, os.linesep))
|
||||
dataToTrafficFile("%s%s" % (responseLogMsg, os.linesep))
|
||||
dataToTrafficFile("%s%s%s%s" % (os.linesep, 76 * '#', os.linesep, os.linesep))
|
||||
|
||||
kb.locks.log.release()
|
||||
with kb.locks.log:
|
||||
dataToTrafficFile("%s%s" % (requestLogMsg, os.linesep))
|
||||
dataToTrafficFile("%s%s" % (responseLogMsg, os.linesep))
|
||||
dataToTrafficFile("%s%s%s%s" % (os.linesep, 76 * '#', os.linesep, os.linesep))
|
||||
|
||||
def getPageTemplate(payload, place):
|
||||
"""
|
||||
|
|
|
@ -42,13 +42,11 @@ class Crawler:
|
|||
threadData = getCurrentThreadData()
|
||||
|
||||
while kb.threadContinue:
|
||||
kb.locks.limits.acquire()
|
||||
if threadData.shared.unprocessed:
|
||||
current = threadData.shared.unprocessed.pop()
|
||||
kb.locks.limits.release()
|
||||
else:
|
||||
kb.locks.limits.release()
|
||||
break
|
||||
with kb.locks.limits:
|
||||
if threadData.shared.unprocessed:
|
||||
current = threadData.shared.unprocessed.pop()
|
||||
else:
|
||||
break
|
||||
|
||||
content = None
|
||||
try:
|
||||
|
@ -83,11 +81,10 @@ class Crawler:
|
|||
continue
|
||||
|
||||
if url.split('.')[-1].lower() not in CRAWL_EXCLUDE_EXTENSIONS:
|
||||
kb.locks.outputs.acquire()
|
||||
threadData.shared.deeper.add(url)
|
||||
if re.search(r"(.*?)\?(.+)", url):
|
||||
threadData.shared.outputs.add(url)
|
||||
kb.locks.outputs.release()
|
||||
with kb.locks.outputs:
|
||||
threadData.shared.deeper.add(url)
|
||||
if re.search(r"(.*?)\?(.+)", url):
|
||||
threadData.shared.outputs.add(url)
|
||||
except UnicodeEncodeError: # for non-HTML files
|
||||
pass
|
||||
finally:
|
||||
|
|
Loading…
Reference in New Issue
Block a user