first implementation of asynchronous inter-protocol communication between the sqlmap RESTful API and the sqlmap engine with SQLite

This commit is contained in:
Bernardo Damele 2013-01-24 12:57:24 +00:00
parent 7d01eb79b4
commit 5b3c8d8991
2 changed files with 47 additions and 43 deletions

View File

@ -13,6 +13,7 @@ import re
import socket import socket
import string import string
import sys import sys
import sqlite3
import threading import threading
import time import time
import urllib2 import urllib2
@ -1793,29 +1794,21 @@ def _mergeOptions(inputOptions, overrideOptions):
if hasattr(conf, key) and conf[key] is None: if hasattr(conf, key) and conf[key] is None:
conf[key] = value conf[key] = value
# Logger recorder object, which keeps the log structure
class LogRecorder(logging.StreamHandler): class LogRecorder(logging.StreamHandler):
"""
Logging handler class which only records CUSTOM_LOGGING.PAYLOAD entries
to a global list.
"""
loghist = []
def emit(self, record): def emit(self, record):
""" """
Simply record the emitted events. Record emitted events to temporary database for asynchronous I/O
communication with the parent process
""" """
self.loghist.append({'levelname': record.levelname, connection = sqlite3.connect(conf.ipc, isolation_level=None)
'text': record.msg % record.args if record.args else record.msg, cursor = connection.cursor()
'id': len(self.loghist) + 1}) cursor.execute("INSERT INTO logs VALUES(NULL, ?, ?, ?)",
(time.strftime("%X"), record.levelname, record.msg % record.args if record.args else record.msg))
if conf.fdLog: cursor.close()
# TODO: this is very heavy operation and slows down a lot the connection.close()
# whole execution of the sqlmap engine, find an alternative
os.write(conf.fdLog, base64pickle(self.loghist))
def _setRestAPILog(): def _setRestAPILog():
if hasattr(conf, "fdLog") and conf.fdLog: if hasattr(conf, "ipc"):
logger.removeHandler(LOGGER_HANDLER) logger.removeHandler(LOGGER_HANDLER)
LOGGER_RECORDER = LogRecorder() LOGGER_RECORDER = LogRecorder()
logger.addHandler(LOGGER_RECORDER) logger.addHandler(LOGGER_RECORDER)

View File

@ -7,6 +7,7 @@ See the file 'doc/COPYING' for copying permission
import os import os
import shutil import shutil
import sqlite3
import tempfile import tempfile
from subprocess import PIPE from subprocess import PIPE
@ -39,7 +40,6 @@ RESTAPI_SERVER_PORT = 8775
# Local global variables # Local global variables
adminid = "" adminid = ""
pipes = dict()
procs = dict() procs = dict()
tasks = AttribDict() tasks = AttribDict()
@ -115,6 +115,19 @@ def task_new():
taskid = hexencode(os.urandom(16)) taskid = hexencode(os.urandom(16))
tasks[taskid] = init_options() tasks[taskid] = init_options()
# Initiate the temporary database for asynchronous I/O with the
# sqlmap engine (children processes)
_, ipc_filepath = tempfile.mkstemp(prefix="sqlmapipc-", suffix=".db", text=False)
connection = sqlite3.connect(ipc_filepath, isolation_level=None)
cursor = connection.cursor()
cursor.execute("DROP TABLE IF EXISTS logs")
cursor.execute("CREATE TABLE logs(id INTEGER PRIMARY KEY AUTOINCREMENT, time TEXT, level TEXT, message TEXT)")
cursor.close()
connection.close()
# Set the temporary database to use for asynchronous I/O communication
tasks[taskid].ipc = ipc_filepath
return jsonize({"taskid": taskid}) return jsonize({"taskid": taskid})
@get("/task/<taskid>/destroy") @get("/task/<taskid>/destroy")
@ -242,7 +255,6 @@ def scan_start(taskid):
""" """
global tasks global tasks
global procs global procs
global pipes
if taskid not in tasks: if taskid not in tasks:
abort(500, "Invalid task ID") abort(500, "Invalid task ID")
@ -253,16 +265,11 @@ def scan_start(taskid):
tasks[taskid][key] = value tasks[taskid][key] = value
# Overwrite output directory (oDir) value to a temporary directory # Overwrite output directory (oDir) value to a temporary directory
tasks[taskid].oDir = tempfile.mkdtemp(prefix="sqlmap-") tasks[taskid].oDir = tempfile.mkdtemp(prefix="sqlmaptask-")
# Launch sqlmap engine in a separate thread # Launch sqlmap engine in a separate thread
logger.debug("starting a scan for task ID %s" % taskid) logger.debug("starting a scan for task ID %s" % taskid)
pipes[taskid] = os.pipe()
# Provide sqlmap engine with the writable pipe for logging
tasks[taskid]["fdLog"] = pipes[taskid][1]
# Launch sqlmap engine # Launch sqlmap engine
procs[taskid] = execute("python sqlmap.py --pickled-options %s" % base64pickle(tasks[taskid]), shell=True, stdin=PIPE, stdout=PIPE, stderr=PIPE, close_fds=False) procs[taskid] = execute("python sqlmap.py --pickled-options %s" % base64pickle(tasks[taskid]), shell=True, stdin=PIPE, stdout=PIPE, stderr=PIPE, close_fds=False)
@ -273,7 +280,6 @@ def scan_output(taskid):
""" """
Read the standard output of sqlmap core execution Read the standard output of sqlmap core execution
""" """
global pipes
global tasks global tasks
if taskid not in tasks: if taskid not in tasks:
@ -303,46 +309,51 @@ def scan_delete(taskid):
@get("/scan/<taskid>/log/<start>/<end>") @get("/scan/<taskid>/log/<start>/<end>")
def scan_log_limited(taskid, start, end): def scan_log_limited(taskid, start, end):
""" """
Retrieve the log messages Retrieve a subset of log messages
""" """
log = None json_log_messages = {}
if taskid not in tasks: if taskid not in tasks:
abort(500, "Invalid task ID") abort(500, "Invalid task ID")
# Temporary "protection" against SQL injection FTW ;)
if not start.isdigit() or not end.isdigit() or end <= start: if not start.isdigit() or not end.isdigit() or end <= start:
abort(500, "Invalid start or end value, must be digits") abort(500, "Invalid start or end value, must be digits")
start = max(0, int(start) - 1) start = max(1, int(start))
end = max(1, int(end)) end = max(1, int(end))
pickledLog = os.read(pipes[taskid][0], 100000)
try: # Read a subset of log messages from the temporary I/O database
log = base64unpickle(pickledLog) connection = sqlite3.connect(tasks[taskid].ipc, isolation_level=None)
log = log[slice(start, end)] cursor = connection.cursor()
except (KeyError, IndexError, TypeError), e: cursor.execute("SELECT id, time, level, message FROM logs WHERE id >= %d AND id <= %d" % (start, end))
logger.error("handled exception when trying to unpickle logger dictionary in scan_log_limited(): %s" % str(e)) db_log_messages = cursor.fetchall()
return jsonize({"log": log}) for (id_, time_, level, message) in db_log_messages:
json_log_messages[id_] = {"time": time_, "level": level, "message": message}
return jsonize({"log": json_log_messages})
@get("/scan/<taskid>/log") @get("/scan/<taskid>/log")
def scan_log(taskid): def scan_log(taskid):
""" """
Retrieve the log messages Retrieve the log messages
""" """
log = None json_log_messages = {}
if taskid not in tasks: if taskid not in tasks:
abort(500, "Invalid task ID") abort(500, "Invalid task ID")
pickledLog = os.read(pipes[taskid][0], 100000) # Read all log messages from the temporary I/O database
connection = sqlite3.connect(tasks[taskid].ipc, isolation_level=None)
cursor = connection.cursor()
cursor.execute("SELECT id, time, level, message FROM logs")
db_log_messages = cursor.fetchall()
try: for (id_, time_, level, message) in db_log_messages:
log = base64unpickle(pickledLog) json_log_messages[id_] = {"time": time_, "level": level, "message": message}
except (KeyError, IndexError, TypeError), e:
logger.error("handled exception when trying to unpickle logger dictionary in scan_log(): %s" % str(e))
return jsonize({"log": log}) return jsonize({"log": json_log_messages})
# Function to handle files inside the output directory # Function to handle files inside the output directory
@get("/download/<taskid>/<target>/<filename:path>") @get("/download/<taskid>/<target>/<filename:path>")