# cython: profile=True from __future__ import unicode_literals from libc.stdlib cimport calloc, free from libcpp.pair cimport pair from cython.operator cimport dereference as deref from murmurhash cimport mrmr from spacy.lexeme cimport Lexeme from spacy.lexeme cimport BLANK_WORD from spacy.string_tools cimport substr from spacy.string_tools cimport to_bytes from spacy.string_tools cimport from_bytes from . import util from os import path cimport cython def get_normalized(unicode lex, size_t length): if lex.isalpha() and lex.islower(): return lex else: return get_word_shape(lex, length) def get_word_shape(lex, length): shape = "" last = "" shape_char = "" seq = 0 for c in lex: if c.isalpha(): if c.isupper(): shape_char = "X" else: shape_char = "x" elif c.isdigit(): shape_char = "d" else: shape_char = c if shape_char == last: seq += 1 else: seq = 0 last = shape_char if seq < 3: shape += shape_char assert shape return shape def set_orth_flags(lex, length): return 0 cdef class Language: def __cinit__(self, name): self.name = name self.bacov = {} self.vocab = new Vocab() self.ortho = new Vocab() self.distri = new Vocab() self.vocab[0].set_empty_key(0) self.distri[0].set_empty_key(0) self.ortho[0].set_empty_key(0) self.vocab[0].set_deleted_key(1) self.distri[0].set_deleted_key(1) self.ortho[0].set_deleted_key(1) self.load_tokenization(util.read_tokenization(name)) cpdef Tokens tokenize(self, unicode string): cdef size_t length = len(string) cdef Py_UNICODE* characters = string cdef size_t i cdef Py_UNICODE c cdef Tokens tokens = Tokens(self) cdef Py_UNICODE* current = calloc(len(string), sizeof(Py_UNICODE)) cdef size_t word_len = 0 cdef Lexeme* token for i in range(length): c = characters[i] if _is_whitespace(c): if word_len != 0: token = self.lookup_chunk(current) while token != NULL: tokens.append(token) token = token.tail for j in range(word_len+1): current[j] = 0 word_len = 0 else: current[word_len] = c word_len += 1 if word_len != 0: token = self.lookup_chunk(current) while token != NULL: tokens.append(token) token = token.tail free(current) return tokens cdef Lexeme_addr lookup(self, unicode string) except 0: cdef size_t length = len(string) if length == 0: return &BLANK_WORD cdef bytes b = to_bytes(string) cdef StringHash hashed = mrmr.hash32(b, len(b) * sizeof(char), 0) # First, check words seen 2+ times cdef Lexeme* word_ptr = self.vocab[0][hashed] if word_ptr == NULL: word_ptr = self.new_lexeme(hashed, string) self.bacov[hashed] = b return word_ptr cdef Lexeme_addr lookup_chunk(self, unicode string) except 0: '''Fetch a Lexeme representing a word string. If the word has not been seen, construct one, splitting off any attached punctuation or clitics. A reference to BLANK_WORD is returned for the empty string. ''' cdef size_t length = len(string) if length == 0: return &BLANK_WORD cdef StringHash hashed = self.hash_string(string) # First, check words seen 2+ times cdef Lexeme* word_ptr = self.vocab[0][hashed] cdef int split if word_ptr == NULL: split = self.find_split(string, length) if split != 0 and split != -1 and split < length: word_ptr = self.new_lexeme(hashed, string[:split]) word_ptr.tail = self.lookup_chunk(string[split:]) else: word_ptr = self.new_lexeme(hashed, string) return word_ptr cdef Orthography* lookup_orth(self, StringHash hashed, unicode lex): cdef Orthography* orth = self.ortho[0][hashed] if orth == NULL: orth = self.new_orth(hashed, lex) return orth cdef Distribution* lookup_dist(self, StringHash hashed): cdef Distribution* dist = self.distri[0][hashed] if dist == NULL: dist = self.new_dist(hashed) return dist cdef Lexeme* new_lexeme(self, StringHash key, unicode string) except NULL: cdef Lexeme* word = calloc(1, sizeof(Lexeme)) word.sic = key word.lex = self.hash_string(string) word.orth = self.lookup_orth(word.lex, string) word.dist = self.lookup_dist(word.lex) self.vocab[0][key] = word return word cdef Orthography* new_orth(self, StringHash hashed, unicode lex) except NULL: cdef unicode last3 cdef unicode norm cdef unicode shape cdef int length length = len(lex) orth = calloc(1, sizeof(Orthography)) orth.first = lex[0] orth.length = length orth.flags = set_orth_flags(lex, orth.length) orth.last3 = self.hash_string(substr(lex, length - 3, length, length)) orth.norm = self.hash_string(get_normalized(lex, length)) orth.shape = self.hash_string(get_word_shape(lex, length)) self.ortho[0][hashed] = orth return orth cdef Distribution* new_dist(self, StringHash hashed) except NULL: dist = calloc(1, sizeof(Distribution)) self.distri[0][hashed] = dist return dist cdef StringHash hash_string(self, unicode s) except 0: '''Hash unicode with MurmurHash64A''' cdef bytes byte_string = to_bytes(s) cdef StringHash hashed = mrmr.hash32(byte_string, len(byte_string) * sizeof(char), 0) self.bacov[hashed] = byte_string return hashed cdef unicode unhash(self, StringHash hash_value): '''Fetch a string from the reverse index, given its hash value.''' return from_bytes(self.bacov[hash_value]) cdef int find_split(self, unicode word, size_t length): return -1 def load_tokenization(self, token_rules=None): cdef Lexeme* word cdef StringHash hashed for chunk, lex, tokens in token_rules: hashed = self.hash_string(chunk) word = self.new_lexeme(hashed, lex) for i, lex in enumerate(tokens): token_string = '%s:@:%d:@:%s' % (chunk, i, lex) length = len(token_string) hashed = self.hash_string(token_string) word.tail = self.new_lexeme(hashed, lex) word = word.tail def load_clusters(self): cdef Lexeme* w data_dir = path.join(path.dirname(__file__), '..', 'data', 'en') case_stats = util.load_case_stats(data_dir) brown_loc = path.join(data_dir, 'clusters') cdef size_t start cdef int end with util.utf8open(brown_loc) as browns_file: for i, line in enumerate(browns_file): cluster_str, token_string, freq_str = line.split() # Decode as a little-endian string, so that we can do & 15 to get # the first 4 bits. See redshift._parse_features.pyx cluster = int(cluster_str[::-1], 2) upper_pc, title_pc = case_stats.get(token_string.lower(), (0.0, 0.0)) hashed = self.hash_string(token_string) word = self.init_lexeme(hashed, token_string) cdef inline bint _is_whitespace(Py_UNICODE c) nogil: if c == ' ': return True elif c == '\n': return True elif c == '\t': return True else: return False cpdef vector[size_t] expand_chunk(size_t addr) except *: cdef vector[size_t] tokens = vector[size_t]() word = addr while word != NULL: tokens.push_back(word) word = word.tail return tokens