mirror of
https://github.com/sqlmapproject/sqlmap.git
synced 2025-01-24 00:04:23 +03:00
Better patch for an Issue #1095
This commit is contained in:
parent
2030311d50
commit
c4c4ac13fe
|
@ -11,12 +11,23 @@ except:
|
||||||
import pickle
|
import pickle
|
||||||
|
|
||||||
import os
|
import os
|
||||||
|
import sys
|
||||||
import tempfile
|
import tempfile
|
||||||
|
|
||||||
from lib.core.exception import SqlmapSystemException
|
from lib.core.exception import SqlmapSystemException
|
||||||
from lib.core.settings import BIGARRAY_CHUNK_LENGTH
|
from lib.core.settings import BIGARRAY_CHUNK_SIZE
|
||||||
from lib.core.settings import BIGARRAY_TEMP_PREFIX
|
from lib.core.settings import BIGARRAY_TEMP_PREFIX
|
||||||
|
|
||||||
|
def _size_of(object):
|
||||||
|
"""
|
||||||
|
Returns total size of a given object (in bytes)
|
||||||
|
"""
|
||||||
|
|
||||||
|
if hasattr(object, "__iter__"):
|
||||||
|
return sum(_size_of(_) for _ in object)
|
||||||
|
else:
|
||||||
|
return sys.getsizeof(object)
|
||||||
|
|
||||||
class Cache(object):
|
class Cache(object):
|
||||||
"""
|
"""
|
||||||
Auxiliary class used for storing cached chunks
|
Auxiliary class used for storing cached chunks
|
||||||
|
@ -34,13 +45,20 @@ class BigArray(list):
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
self.chunks = [[]]
|
self.chunks = [[]]
|
||||||
|
self.chunk_length = sys.maxint
|
||||||
self.cache = None
|
self.cache = None
|
||||||
self.filenames = set()
|
self.filenames = set()
|
||||||
self._os_remove = os.remove
|
self._os_remove = os.remove
|
||||||
|
self._size_counter = 0
|
||||||
|
|
||||||
def append(self, value):
|
def append(self, value):
|
||||||
self.chunks[-1].append(value)
|
self.chunks[-1].append(value)
|
||||||
if len(self.chunks[-1]) >= BIGARRAY_CHUNK_LENGTH:
|
if self.chunk_length == sys.maxint:
|
||||||
|
self._size_counter += _size_of(value)
|
||||||
|
if self._size_counter >= BIGARRAY_CHUNK_SIZE:
|
||||||
|
self.chunk_length = len(self.chunks[-1])
|
||||||
|
self._size_counter = None
|
||||||
|
if len(self.chunks[-1]) >= self.chunk_length:
|
||||||
filename = self._dump(self.chunks[-1])
|
filename = self._dump(self.chunks[-1])
|
||||||
self.chunks[-1] = filename
|
self.chunks[-1] = filename
|
||||||
self.chunks.append([])
|
self.chunks.append([])
|
||||||
|
@ -54,7 +72,7 @@ class BigArray(list):
|
||||||
self.chunks.pop()
|
self.chunks.pop()
|
||||||
try:
|
try:
|
||||||
with open(self.chunks[-1], "rb") as fp:
|
with open(self.chunks[-1], "rb") as fp:
|
||||||
self.chunks[-1] = self._load(fp)
|
self.chunks[-1] = pickle.load(fp)
|
||||||
except IOError, ex:
|
except IOError, ex:
|
||||||
errMsg = "exception occurred while retrieving data "
|
errMsg = "exception occurred while retrieving data "
|
||||||
errMsg += "from a temporary file ('%s')" % ex
|
errMsg += "from a temporary file ('%s')" % ex
|
||||||
|
@ -67,35 +85,13 @@ class BigArray(list):
|
||||||
return index
|
return index
|
||||||
return ValueError, "%s is not in list" % value
|
return ValueError, "%s is not in list" % value
|
||||||
|
|
||||||
def _load(self, fp):
|
|
||||||
retval = []
|
|
||||||
unpickler = pickle.Unpickler(fp)
|
|
||||||
_ = unpickler.load()
|
|
||||||
if not isinstance(_, list):
|
|
||||||
retval.append(_)
|
|
||||||
while True:
|
|
||||||
try:
|
|
||||||
retval.append(unpickler.load())
|
|
||||||
except EOFError:
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
retval = _
|
|
||||||
return retval
|
|
||||||
|
|
||||||
def _dump(self, chunk):
|
def _dump(self, chunk):
|
||||||
try:
|
try:
|
||||||
handle, filename = tempfile.mkstemp(prefix=BIGARRAY_TEMP_PREFIX)
|
handle, filename = tempfile.mkstemp(prefix=BIGARRAY_TEMP_PREFIX)
|
||||||
self.filenames.add(filename)
|
self.filenames.add(filename)
|
||||||
os.close(handle)
|
os.close(handle)
|
||||||
try:
|
with open(filename, "w+b") as fp:
|
||||||
with open(filename, "w+b") as fp:
|
pickle.dump(chunk, fp, pickle.HIGHEST_PROTOCOL)
|
||||||
pickle.dump(chunk, fp, pickle.HIGHEST_PROTOCOL)
|
|
||||||
except MemoryError:
|
|
||||||
with open(filename, "w+b") as fp:
|
|
||||||
pickler = pickle.Pickler(fp, pickle.HIGHEST_PROTOCOL)
|
|
||||||
pickler.fast = True
|
|
||||||
for value in chunk:
|
|
||||||
pickler.dump(value)
|
|
||||||
return filename
|
return filename
|
||||||
except IOError, ex:
|
except IOError, ex:
|
||||||
errMsg = "exception occurred while storing data "
|
errMsg = "exception occurred while storing data "
|
||||||
|
@ -109,7 +105,7 @@ class BigArray(list):
|
||||||
if not (self.cache and self.cache.index == index):
|
if not (self.cache and self.cache.index == index):
|
||||||
try:
|
try:
|
||||||
with open(self.chunks[index], "rb") as fp:
|
with open(self.chunks[index], "rb") as fp:
|
||||||
self.cache = Cache(index, self._load(fp), False)
|
self.cache = Cache(index, pickle.load(fp), False)
|
||||||
except IOError, ex:
|
except IOError, ex:
|
||||||
errMsg = "exception occurred while retrieving data "
|
errMsg = "exception occurred while retrieving data "
|
||||||
errMsg += "from a temporary file ('%s')" % ex
|
errMsg += "from a temporary file ('%s')" % ex
|
||||||
|
@ -133,8 +129,8 @@ class BigArray(list):
|
||||||
def __getitem__(self, y):
|
def __getitem__(self, y):
|
||||||
if y < 0:
|
if y < 0:
|
||||||
y += len(self)
|
y += len(self)
|
||||||
index = y / BIGARRAY_CHUNK_LENGTH
|
index = y / self.chunk_length
|
||||||
offset = y % BIGARRAY_CHUNK_LENGTH
|
offset = y % self.chunk_length
|
||||||
chunk = self.chunks[index]
|
chunk = self.chunks[index]
|
||||||
if isinstance(chunk, list):
|
if isinstance(chunk, list):
|
||||||
return chunk[offset]
|
return chunk[offset]
|
||||||
|
@ -143,8 +139,8 @@ class BigArray(list):
|
||||||
return self.cache.data[offset]
|
return self.cache.data[offset]
|
||||||
|
|
||||||
def __setitem__(self, y, value):
|
def __setitem__(self, y, value):
|
||||||
index = y / BIGARRAY_CHUNK_LENGTH
|
index = y / self.chunk_length
|
||||||
offset = y % BIGARRAY_CHUNK_LENGTH
|
offset = y % self.chunk_length
|
||||||
chunk = self.chunks[index]
|
chunk = self.chunks[index]
|
||||||
if isinstance(chunk, list):
|
if isinstance(chunk, list):
|
||||||
chunk[offset] = value
|
chunk[offset] = value
|
||||||
|
@ -161,4 +157,4 @@ class BigArray(list):
|
||||||
yield self[i]
|
yield self[i]
|
||||||
|
|
||||||
def __len__(self):
|
def __len__(self):
|
||||||
return len(self.chunks[-1]) if len(self.chunks) == 1 else (len(self.chunks) - 1) * BIGARRAY_CHUNK_LENGTH + len(self.chunks[-1])
|
return len(self.chunks[-1]) if len(self.chunks) == 1 else (len(self.chunks) - 1) * self.chunk_length + len(self.chunks[-1])
|
||||||
|
|
|
@ -443,8 +443,8 @@ WAF_ATTACK_VECTORS = (
|
||||||
# Used for status representation in dictionary attack phase
|
# Used for status representation in dictionary attack phase
|
||||||
ROTATING_CHARS = ('\\', '|', '|', '/', '-')
|
ROTATING_CHARS = ('\\', '|', '|', '/', '-')
|
||||||
|
|
||||||
# Chunk length (in items) used by BigArray objects (only last chunk and cached one are held in memory)
|
# Approximate chunk length (in bytes) used by BigArray objects (only last chunk and cached one are held in memory)
|
||||||
BIGARRAY_CHUNK_LENGTH = 1024
|
BIGARRAY_CHUNK_SIZE = 1024 * 1024
|
||||||
|
|
||||||
# Prefix used for storing dumped chunks in BigArray objects
|
# Prefix used for storing dumped chunks in BigArray objects
|
||||||
BIGARRAY_TEMP_PREFIX = "sqlmapba-%d-" % os.getpid()
|
BIGARRAY_TEMP_PREFIX = "sqlmapba-%d-" % os.getpid()
|
||||||
|
|
Loading…
Reference in New Issue
Block a user