mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-25 17:36:30 +03:00
194 lines
6.1 KiB
Cython
194 lines
6.1 KiB
Cython
from __future__ import unicode_literals
|
|
import codecs
|
|
|
|
from libc.string cimport memcpy
|
|
from murmurhash.mrmr cimport hash64
|
|
|
|
from preshed.maps cimport map_iter, key_t
|
|
|
|
from cpython cimport PyUnicode_AS_DATA
|
|
from cpython cimport PyUnicode_GET_DATA_SIZE
|
|
|
|
from libc.stdint cimport int64_t
|
|
|
|
|
|
from .typedefs cimport hash_t, attr_t
|
|
|
|
try:
|
|
import codecs as io
|
|
except ImportError:
|
|
import io
|
|
|
|
import ujson as json
|
|
|
|
|
|
cpdef hash_t hash_string(unicode string) except 0:
|
|
# This has to be like this for
|
|
chars = <char*>PyUnicode_AS_DATA(string)
|
|
size = PyUnicode_GET_DATA_SIZE(string)
|
|
return hash64(chars, size, 1)
|
|
|
|
|
|
cdef unicode _decode(const Utf8Str* string):
|
|
cdef int i, length
|
|
if string.s[0] < sizeof(string.s) and string.s[0] != 0:
|
|
return string.s[1:string.s[0]+1].decode('utf8')
|
|
elif string.p[0] < 255:
|
|
return string.p[1:string.p[0]+1].decode('utf8')
|
|
else:
|
|
i = 0
|
|
length = 0
|
|
while string.p[i] == 255:
|
|
i += 1
|
|
length += 255
|
|
length += string.p[i]
|
|
i += 1
|
|
return string.p[i:length + i].decode('utf8')
|
|
|
|
|
|
cdef Utf8Str _allocate(Pool mem, const unsigned char* chars, int length) except *:
|
|
cdef int n_length_bytes
|
|
cdef int i
|
|
cdef Utf8Str string
|
|
assert length != 0
|
|
if length < sizeof(string.s):
|
|
string.s[0] = <unsigned char>length
|
|
memcpy(&string.s[1], chars, length)
|
|
return string
|
|
elif length < 255:
|
|
string.p = <unsigned char*>mem.alloc(length + 1, sizeof(unsigned char))
|
|
string.p[0] = length
|
|
memcpy(&string.p[1], chars, length)
|
|
assert string.s[0] >= sizeof(string.s) or string.s[0] == 0, string.s[0]
|
|
return string
|
|
else:
|
|
i = 0
|
|
n_length_bytes = (length // 255) + 1
|
|
string.p = <unsigned char*>mem.alloc(length + n_length_bytes, sizeof(unsigned char))
|
|
for i in range(n_length_bytes-1):
|
|
string.p[i] = 255
|
|
string.p[n_length_bytes-1] = length % 255
|
|
memcpy(&string.p[n_length_bytes], chars, length)
|
|
assert string.s[0] >= sizeof(string.s) or string.s[0] == 0, string.s[0]
|
|
return string
|
|
|
|
|
|
cdef class StringStore:
|
|
'''Map strings to and from integer IDs.'''
|
|
def __init__(self, strings=None):
|
|
self.mem = Pool()
|
|
self._map = PreshMap()
|
|
self._resize_at = 10000
|
|
self.c = <Utf8Str*>self.mem.alloc(self._resize_at, sizeof(Utf8Str))
|
|
self.size = 1
|
|
if strings is not None:
|
|
for string in strings:
|
|
_ = self[string]
|
|
|
|
property size:
|
|
def __get__(self):
|
|
return self.size -1
|
|
|
|
def __len__(self):
|
|
return self.size-1
|
|
|
|
def __getitem__(self, object string_or_id):
|
|
cdef bytes byte_string
|
|
cdef unicode py_string
|
|
cdef const Utf8Str* utf8str
|
|
|
|
cdef int id_
|
|
if isinstance(string_or_id, int) or isinstance(string_or_id, long):
|
|
if string_or_id == 0:
|
|
return u''
|
|
elif string_or_id < 1 or string_or_id >= self.size:
|
|
raise IndexError(string_or_id)
|
|
utf8str = &self.c[<int>string_or_id]
|
|
return _decode(utf8str)
|
|
elif isinstance(string_or_id, bytes):
|
|
if len(string_or_id) == 0:
|
|
return 0
|
|
py_string = string_or_id.decode('utf8')
|
|
utf8str = self.intern(py_string)
|
|
return utf8str - self.c
|
|
elif isinstance(string_or_id, unicode):
|
|
if len(string_or_id) == 0:
|
|
return 0
|
|
py_string = string_or_id
|
|
utf8str = self.intern(py_string)
|
|
return utf8str - self.c
|
|
else:
|
|
raise TypeError(type(string_or_id))
|
|
|
|
def __iter__(self):
|
|
cdef int i
|
|
for i in range(self.size):
|
|
if i == 0:
|
|
yield u''
|
|
else:
|
|
utf8str = &self.c[i]
|
|
yield _decode(utf8str)
|
|
|
|
def __reduce__(self):
|
|
strings = [""]
|
|
for i in range(1, self.size):
|
|
string = &self.c[i]
|
|
py_string = _decode(string)
|
|
strings.append(py_string)
|
|
return (StringStore, (strings,), None, None, None)
|
|
|
|
cdef const Utf8Str* intern(self, unicode py_string) except NULL:
|
|
# 0 means missing, but we don't bother offsetting the index.
|
|
cdef hash_t key = hash_string(py_string)
|
|
value = <Utf8Str*>self._map.get(key)
|
|
if value != NULL:
|
|
return value
|
|
|
|
if self.size == self._resize_at:
|
|
self._realloc()
|
|
cdef bytes byte_string = py_string.encode('utf8')
|
|
self.c[self.size] = _allocate(self.mem, <unsigned char*>byte_string, len(byte_string))
|
|
self._map.set(key, <void*>&self.c[self.size])
|
|
self.size += 1
|
|
return &self.c[self.size-1]
|
|
|
|
def dump(self, file_):
|
|
string_data = json.dumps([s for s in self])
|
|
if not isinstance(string_data, unicode):
|
|
string_data = string_data.decode('utf8')
|
|
file_.write(string_data)
|
|
|
|
def load(self, file_):
|
|
strings = json.load(file_)
|
|
if strings == ['']:
|
|
return None
|
|
cdef unicode string
|
|
for string in strings:
|
|
if string:
|
|
self.intern(string)
|
|
|
|
def _realloc(self):
|
|
# We want to map straight to pointers, but they'll be invalidated if
|
|
# we resize our array. So, first we remap to indices, then we resize,
|
|
# then we can acquire the new pointers.
|
|
cdef Pool tmp_mem = Pool()
|
|
keys = <key_t*>tmp_mem.alloc(self.size, sizeof(key_t))
|
|
cdef key_t key
|
|
cdef void* value
|
|
cdef const Utf8Str ptr
|
|
cdef int i = 0
|
|
cdef size_t offset
|
|
while map_iter(self._map.c_map, &i, &key, &value):
|
|
# Find array index with pointer arithmetic
|
|
offset = ((<Utf8Str*>value) - self.c)
|
|
keys[offset] = key
|
|
|
|
self._resize_at *= 2
|
|
cdef size_t new_size = self._resize_at * sizeof(Utf8Str)
|
|
self.c = <Utf8Str*>self.mem.realloc(self.c, new_size)
|
|
|
|
self._map = PreshMap(self.size)
|
|
for i in range(self.size):
|
|
if keys[i]:
|
|
self._map.set(keys[i], &self.c[i])
|