mirror of
https://github.com/sqlmapproject/sqlmap.git
synced 2025-02-09 08:00:36 +03:00
minor update
This commit is contained in:
parent
07e2c72943
commit
cda39ca350
|
@ -394,7 +394,17 @@ def __setCrawler():
|
||||||
return
|
return
|
||||||
|
|
||||||
crawler = Crawler()
|
crawler = Crawler()
|
||||||
crawler.getTargetUrls()
|
depth = 1
|
||||||
|
|
||||||
|
infoMsg = "setting crawling options"
|
||||||
|
logger.info(infoMsg)
|
||||||
|
|
||||||
|
message = "please enter maximum depth [Enter for %d (default)] " % depth
|
||||||
|
choice = readInput(message, default=str(depth))
|
||||||
|
if choice and choice.isdigit():
|
||||||
|
depth = int(choice)
|
||||||
|
|
||||||
|
crawler.getTargetUrls(depth)
|
||||||
|
|
||||||
def __setGoogleDorking():
|
def __setGoogleDorking():
|
||||||
"""
|
"""
|
||||||
|
|
|
@ -70,7 +70,7 @@ class Crawler:
|
||||||
threadData.shared.deeper = set()
|
threadData.shared.deeper = set()
|
||||||
threadData.shared.unprocessed = set([conf.url])
|
threadData.shared.unprocessed = set([conf.url])
|
||||||
|
|
||||||
logger.info("starting crawling")
|
logger.info("starting crawler")
|
||||||
|
|
||||||
for i in xrange(depth):
|
for i in xrange(depth):
|
||||||
numThreads = min(conf.threads, len(threadData.shared.unprocessed))
|
numThreads = min(conf.threads, len(threadData.shared.unprocessed))
|
||||||
|
|
Loading…
Reference in New Issue
Block a user