mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 01:48:04 +03:00 
			
		
		
		
	* Remove the feature array stuff from Tokens class, and replace vector with array-based implementation, with padding.
This commit is contained in:
		
							parent
							
								
									ea1d4a81eb
								
							
						
					
					
						commit
						e5e951ae67
					
				| 
						 | 
				
			
			@ -59,5 +59,5 @@ cdef class Language:
 | 
			
		|||
    cdef int _find_prefix(self, Py_UNICODE* characters, size_t length) except -1
 | 
			
		||||
    cdef int _find_suffix(self, Py_UNICODE* characters, size_t length) except -1
 | 
			
		||||
    cdef int _find_infix(self, Py_UNICODE* characters, size_t length) except -1
 | 
			
		||||
    cdef int _save_cached(self, vector[LexemeC*] *tokens, uint64_t key, size_t n) except -1
 | 
			
		||||
    cdef int _save_cached(self, LexemeC** tokens, uint64_t key, int n) except -1
 | 
			
		||||
 
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -101,10 +101,10 @@ cdef class Language:
 | 
			
		|||
        cdef uint64_t orig_key
 | 
			
		||||
        cdef int orig_size
 | 
			
		||||
        orig_key = span.key
 | 
			
		||||
        orig_size = tokens.lex.size()
 | 
			
		||||
        orig_size = tokens.length
 | 
			
		||||
        self._split_affixes(span, &prefixes, &suffixes)
 | 
			
		||||
        self._attach_tokens(tokens, start, span, &prefixes, &suffixes)
 | 
			
		||||
        self._save_cached(tokens.lex, orig_key, orig_size)
 | 
			
		||||
        self._save_cached(&tokens.lex[orig_size], orig_key, tokens.length - orig_size)
 | 
			
		||||
 | 
			
		||||
    cdef String* _split_affixes(self, String* string, vector[LexemeC*] *prefixes,
 | 
			
		||||
                                vector[LexemeC*] *suffixes) except NULL:
 | 
			
		||||
| 
						 | 
				
			
			@ -177,12 +177,11 @@ cdef class Language:
 | 
			
		|||
            idx = tokens.push_back(idx, deref(it))
 | 
			
		||||
            preinc(it)
 | 
			
		||||
 | 
			
		||||
    cdef int _save_cached(self, vector[LexemeC*] *tokens, uint64_t key, size_t n) except -1:
 | 
			
		||||
        assert tokens.size() > n
 | 
			
		||||
        lexemes = <LexemeC**>self._mem.alloc((tokens.size() - n) + 1, sizeof(LexemeC**))
 | 
			
		||||
        cdef size_t i, j
 | 
			
		||||
        for i, j in enumerate(range(n, tokens.size())):
 | 
			
		||||
            lexemes[i] = tokens.at(j)
 | 
			
		||||
    cdef int _save_cached(self, LexemeC** tokens, uint64_t key, int n) except -1:
 | 
			
		||||
        lexemes = <LexemeC**>self._mem.alloc(n + 1, sizeof(LexemeC**))
 | 
			
		||||
        cdef int i
 | 
			
		||||
        for i in range(n):
 | 
			
		||||
            lexemes[i] = tokens[i]
 | 
			
		||||
        lexemes[i + 1] = NULL
 | 
			
		||||
        self.cache.set(key, lexemes)
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -73,6 +73,9 @@ cdef struct LexemeC:
 | 
			
		|||
    flag_t dist_flags
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
cdef LexemeC EMPTY_LEXEME
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
cpdef dict get_lexeme_dict(size_t i, unicode string)
 | 
			
		||||
 | 
			
		||||
cdef char* intern_and_encode(unicode string, size_t* length) except NULL
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -1,10 +1,13 @@
 | 
			
		|||
from cpython.ref cimport Py_INCREF
 | 
			
		||||
from cymem.cymem cimport Pool
 | 
			
		||||
 | 
			
		||||
from libc.string cimport memset
 | 
			
		||||
 | 
			
		||||
import orth
 | 
			
		||||
 | 
			
		||||
OOV_DIST_FLAGS = 0
 | 
			
		||||
 | 
			
		||||
memset(&EMPTY_LEXEME, 0, sizeof(LexemeC))
 | 
			
		||||
 | 
			
		||||
cpdef dict get_lexeme_dict(size_t i, unicode string):
 | 
			
		||||
    ints = [None for _ in range(LexInt_N)]
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -1,21 +1,24 @@
 | 
			
		|||
from cymem.cymem cimport Pool
 | 
			
		||||
 | 
			
		||||
from spacy.lexeme cimport LexemeC
 | 
			
		||||
from libcpp.vector cimport vector
 | 
			
		||||
from thinc.typedefs cimport atom_t
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
cdef class Tokens:
 | 
			
		||||
    cdef vector[LexemeC*] *lex
 | 
			
		||||
    cdef vector[int] *idx
 | 
			
		||||
    cdef vector[int] *pos
 | 
			
		||||
    cdef Pool mem
 | 
			
		||||
 | 
			
		||||
    cdef LexemeC** _lex_ptr
 | 
			
		||||
    cdef int* _idx_ptr
 | 
			
		||||
    cdef int* _pos_ptr
 | 
			
		||||
    cdef LexemeC** lex
 | 
			
		||||
    cdef int* idx
 | 
			
		||||
    cdef int* pos
 | 
			
		||||
 | 
			
		||||
    cdef int length
 | 
			
		||||
    cdef int max_length
 | 
			
		||||
 | 
			
		||||
    cdef int extend(self, int i, LexemeC** lexemes, int n) except -1
 | 
			
		||||
    cdef int push_back(self, int i, LexemeC* lexeme) except -1
 | 
			
		||||
    cdef int int_array(self, atom_t* atoms, int i, int* indices, int n_idx,
 | 
			
		||||
                       int* features, int n_feat) except -1
 | 
			
		||||
    cdef int string_array(self, atom_t* atoms, int i, int* indices, int n_idx,
 | 
			
		||||
                          int* features, int n_feat) except -1
 | 
			
		||||
    cdef int bool_array(self, atom_t* atoms, int i, int* indices, int n_idx,
 | 
			
		||||
                        int* features, int n_feat) except -1
 | 
			
		||||
 | 
			
		||||
    cpdef int id(self, size_t i) except -1
 | 
			
		||||
    cpdef float prob(self, size_t i) except 1
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
							
								
								
									
										221
									
								
								spacy/tokens.pyx
									
									
									
									
									
								
							
							
						
						
									
										221
									
								
								spacy/tokens.pyx
									
									
									
									
									
								
							| 
						 | 
				
			
			@ -6,6 +6,14 @@ cimport numpy
 | 
			
		|||
cimport cython
 | 
			
		||||
import numpy
 | 
			
		||||
 | 
			
		||||
DEF PADDING = 5
 | 
			
		||||
 | 
			
		||||
cdef int bounds_check(int i, int length, int padding) except -1:
 | 
			
		||||
    if (i + padding) < 0:
 | 
			
		||||
        raise IndexError
 | 
			
		||||
    if (i - padding) >= length:
 | 
			
		||||
        raise IndexError
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
cdef class Tokens:
 | 
			
		||||
    """A sequence of references to Lexeme objects.
 | 
			
		||||
| 
						 | 
				
			
			@ -26,71 +34,58 @@ cdef class Tokens:
 | 
			
		|||
    >>> tokens.can_noun(1)
 | 
			
		||||
    True
 | 
			
		||||
    """
 | 
			
		||||
    def __cinit__(self, string_length=0):
 | 
			
		||||
        size = int(string_length / 3) if string_length >= 3 else 1
 | 
			
		||||
        self.lex = new vector[LexemeC*]()
 | 
			
		||||
        self.idx = new vector[int]()
 | 
			
		||||
        self.pos = new vector[int]()
 | 
			
		||||
        self.lex.reserve(size)
 | 
			
		||||
        self.idx.reserve(size)
 | 
			
		||||
        self.pos.reserve(size)
 | 
			
		||||
    def __init__(self, string_length=0):
 | 
			
		||||
        if string_length >= 3:
 | 
			
		||||
            size = int(string_length / 3.0)
 | 
			
		||||
        else:
 | 
			
		||||
            size = 5
 | 
			
		||||
        self.mem = Pool()
 | 
			
		||||
        # Guarantee self.lex[i-x], for any i >= 0 and x < padding is in bounds
 | 
			
		||||
        # However, we need to remember the true starting places, so that we can
 | 
			
		||||
        # realloc.
 | 
			
		||||
        self._lex_ptr = <LexemeC**>self.mem.alloc(size + (PADDING*2), sizeof(LexemeC*))
 | 
			
		||||
        self._idx_ptr = <int*>self.mem.alloc(size + (PADDING*2), sizeof(int))
 | 
			
		||||
        self._pos_ptr = <int*>self.mem.alloc(size + (PADDING*2), sizeof(int))
 | 
			
		||||
        self.lex = self._lex_ptr
 | 
			
		||||
        self.idx = self._idx_ptr
 | 
			
		||||
        self.pos = self._pos_ptr
 | 
			
		||||
        for i in range(PADDING):
 | 
			
		||||
            self.lex[i] = &EMPTY_LEXEME
 | 
			
		||||
        for i in range(size, PADDING):
 | 
			
		||||
            self.lex[i] = &EMPTY_LEXEME
 | 
			
		||||
        self.lex += PADDING
 | 
			
		||||
        self.idx += PADDING
 | 
			
		||||
        self.pos += PADDING
 | 
			
		||||
 | 
			
		||||
    def __dealloc__(self):
 | 
			
		||||
        del self.lex
 | 
			
		||||
        del self.idx
 | 
			
		||||
        del self.pos
 | 
			
		||||
        self.max_length = size
 | 
			
		||||
        self.length = 0
 | 
			
		||||
 | 
			
		||||
    def __getitem__(self, i):
 | 
			
		||||
        if i >= self.lex.size():
 | 
			
		||||
            raise IndexError
 | 
			
		||||
        return Lexeme(<size_t>self.lex.at(i))
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return Lexeme(<size_t>self.lex[i])
 | 
			
		||||
 | 
			
		||||
    def __len__(self):
 | 
			
		||||
        return self.lex.size()
 | 
			
		||||
        return self.length
 | 
			
		||||
 | 
			
		||||
    cdef int push_back(self, int idx, LexemeC* lexeme) except -1:
 | 
			
		||||
        self.lex.push_back(lexeme)
 | 
			
		||||
        self.idx.push_back(idx)
 | 
			
		||||
        if self.length == self.max_length:
 | 
			
		||||
            self._realloc(self.length * 2)
 | 
			
		||||
        self.lex[self.length] = lexeme
 | 
			
		||||
        self.idx[self.length] = idx
 | 
			
		||||
        self.pos[self.length] = 0
 | 
			
		||||
        self.length += 1
 | 
			
		||||
        return idx + lexeme.ints[<int>LexInt_length]
 | 
			
		||||
 | 
			
		||||
    cdef int int_array(self, atom_t* output, int i, int* indices, int n_idx,
 | 
			
		||||
                       int* features, int n_feat):
 | 
			
		||||
        cdef int feat_id, idx
 | 
			
		||||
        cdef int length = self.lex.size()
 | 
			
		||||
        for feat_id in features[:n_feat]:
 | 
			
		||||
            for idx in indices[:n_idx]:
 | 
			
		||||
                if idx < 0 or idx >= length:
 | 
			
		||||
                    output[i] = 0
 | 
			
		||||
                else:
 | 
			
		||||
                    output[i] = self.lex[0][idx].ints[<int>feat_id]
 | 
			
		||||
                i += 1
 | 
			
		||||
        return i
 | 
			
		||||
    def _realloc(self, new_size):
 | 
			
		||||
        self.max_length = new_size
 | 
			
		||||
        n = new_size + (PADDING * 2)
 | 
			
		||||
        self._lex_ptr = <LexemeC**>self.mem.realloc(self._lex_ptr, n * sizeof(LexemeC*))
 | 
			
		||||
        self._idx_ptr = <int*>self.mem.realloc(self._idx_ptr, n * sizeof(int))
 | 
			
		||||
        self._pos_ptr = <int*>self.mem.realloc(self._pos_ptr, n * sizeof(int))
 | 
			
		||||
        self.lex = self._lex_ptr + PADDING
 | 
			
		||||
        self.idx = self._idx_ptr + PADDING
 | 
			
		||||
        self.pos = self._pos_ptr + PADDING
 | 
			
		||||
 | 
			
		||||
    cdef int string_array(self, atom_t* output, int i, int* indices, int n_idx,
 | 
			
		||||
                          int* features, int n_feat):
 | 
			
		||||
        cdef int feat_id, idx
 | 
			
		||||
        cdef int length = self.lex.size()
 | 
			
		||||
        for feat_id in features[:n_feat]:
 | 
			
		||||
            for idx in indices[:n_idx]:
 | 
			
		||||
                if idx < 0 or idx >= length:
 | 
			
		||||
                    output[i] = 0
 | 
			
		||||
                else:
 | 
			
		||||
                    output[i] = <atom_t>self.lex[0][idx].strings[<int>feat_id]
 | 
			
		||||
                i += 1
 | 
			
		||||
        return i
 | 
			
		||||
 | 
			
		||||
    cdef int bool_array(self, atom_t* output, int i, int* indices, int n_idx,
 | 
			
		||||
                        int* features, int n_feat):
 | 
			
		||||
        cdef int feat_id, idx
 | 
			
		||||
        cdef int length = self.lex.size()
 | 
			
		||||
        for feat_id in features[:n_feat]:
 | 
			
		||||
            for idx in indices[:n_idx]:
 | 
			
		||||
                if idx < 0 or idx >= length:
 | 
			
		||||
                    output[i] = 0
 | 
			
		||||
                else:
 | 
			
		||||
                    output[i] = lexeme_check_dist_flag(self.lex[0][idx], feat_id)
 | 
			
		||||
                i += 1
 | 
			
		||||
        return i
 | 
			
		||||
 | 
			
		||||
    cdef int extend(self, int idx, LexemeC** lexemes, int n) except -1:
 | 
			
		||||
        cdef int i
 | 
			
		||||
| 
						 | 
				
			
			@ -99,131 +94,161 @@ cdef class Tokens:
 | 
			
		|||
        elif n == 0:
 | 
			
		||||
            i = 0
 | 
			
		||||
            while lexemes[i] != NULL:
 | 
			
		||||
                self.lex.push_back(lexemes[i])
 | 
			
		||||
                self.idx.push_back(idx)
 | 
			
		||||
                idx += lexemes[i].ints[<int>LexInt_length]
 | 
			
		||||
                idx = self.push_back(idx, lexemes[i])
 | 
			
		||||
                i += 1
 | 
			
		||||
        else:
 | 
			
		||||
            for i in range(n):
 | 
			
		||||
                self.lex.push_back(lexemes[i])
 | 
			
		||||
                self.idx.push_back(idx)
 | 
			
		||||
                idx += lexemes[i].ints[<int>LexInt_length]
 | 
			
		||||
                idx = self.push_back(idx, lexemes[i])
 | 
			
		||||
        return idx
 | 
			
		||||
 | 
			
		||||
    cpdef int id(self, size_t i) except -1:
 | 
			
		||||
        return self.lex.at(i).ints[<int>LexInt_id]
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return self.lex[i].ints[<int>LexInt_id]
 | 
			
		||||
 | 
			
		||||
    cpdef float prob(self, size_t i) except 1:
 | 
			
		||||
        return self.lex.at(i).floats[<int>LexFloat_prob]
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return self.lex[i].floats[<int>LexFloat_prob]
 | 
			
		||||
 | 
			
		||||
    cpdef int cluster(self, size_t i) except *:
 | 
			
		||||
        return self.lex.at(i).ints[<int>LexInt_cluster]
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return self.lex[i].ints[<int>LexInt_cluster]
 | 
			
		||||
 | 
			
		||||
    cpdef bint check_orth_flag(self, size_t i, size_t flag_id) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), flag_id)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], flag_id)
 | 
			
		||||
 | 
			
		||||
    cpdef bint check_dist_flag(self, size_t i, size_t flag_id) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), flag_id)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], flag_id)
 | 
			
		||||
 | 
			
		||||
    cpdef unicode string_view(self, size_t i, size_t view_id):
 | 
			
		||||
        return lexeme_get_string(self.lex.at(i), view_id)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_get_string(self.lex[i], view_id)
 | 
			
		||||
 | 
			
		||||
    # Provide accessor methods for the features supported by the language.
 | 
			
		||||
    # Without these, clients have to use the underlying string_view and check_flag
 | 
			
		||||
    # methods, which requires them to know the IDs.
 | 
			
		||||
 | 
			
		||||
    cpdef unicode string(self, size_t i):
 | 
			
		||||
        if i >= self.lex.size():
 | 
			
		||||
            raise IndexError
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return self.orig(i)
 | 
			
		||||
 | 
			
		||||
    cpdef unicode orig(self, size_t i):
 | 
			
		||||
        cdef bytes utf8_string = self.lex.at(i).strings[<int>LexStr_orig]
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        cdef bytes utf8_string = self.lex[i].strings[<int>LexStr_orig]
 | 
			
		||||
        cdef unicode string = utf8_string.decode('utf8')
 | 
			
		||||
        return string
 | 
			
		||||
 | 
			
		||||
    cpdef unicode norm(self, size_t i):
 | 
			
		||||
        cdef bytes utf8_string = self.lex.at(i).strings[<int>LexStr_norm]
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        cdef bytes utf8_string = self.lex[i].strings[<int>LexStr_norm]
 | 
			
		||||
        cdef unicode string = utf8_string.decode('utf8')
 | 
			
		||||
        return string
 | 
			
		||||
 | 
			
		||||
    cpdef unicode shape(self, size_t i):
 | 
			
		||||
        return lexeme_get_string(self.lex.at(i), LexStr_shape)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_get_string(self.lex[i], LexStr_shape)
 | 
			
		||||
 | 
			
		||||
    cpdef unicode unsparse(self, size_t i):
 | 
			
		||||
        return lexeme_get_string(self.lex.at(i), LexStr_unsparse)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_get_string(self.lex[i], LexStr_unsparse)
 | 
			
		||||
 | 
			
		||||
    cpdef unicode asciied(self, size_t i):
 | 
			
		||||
        return lexeme_get_string(self.lex.at(i), LexStr_asciied)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_get_string(self.lex[i], LexStr_asciied)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_alpha(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_alpha)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_alpha)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_ascii(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_ascii)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_ascii)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_digit(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_digit)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_digit)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_lower(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_lower)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_lower)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_punct(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_punct)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_punct)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_space(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_space)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_space)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_title(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_title)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_title)
 | 
			
		||||
 | 
			
		||||
    cpdef bint is_upper(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex.at(i), LexOrth_upper)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_orth_flag(self.lex[i], LexOrth_upper)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_adj(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_adj)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_adj)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_adp(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_adp)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_adp)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_adv(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_adv)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_adv)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_conj(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_conj)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_conj)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_det(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_det)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_det)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_noun(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_noun)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_noun)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_num(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_num)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_num)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_pdt(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_pdt)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_pdt)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_pos(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_pos)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_pos)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_pron(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_pron)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_pron)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_prt(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_prt)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_prt)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_punct(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_punct)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_punct)
 | 
			
		||||
 | 
			
		||||
    cpdef bint can_verb(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_verb)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_verb)
 | 
			
		||||
 | 
			
		||||
    cpdef bint oft_lower(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_lower)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_lower)
 | 
			
		||||
 | 
			
		||||
    cpdef bint oft_title(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_title)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_title)
 | 
			
		||||
 | 
			
		||||
    cpdef bint oft_upper(self, size_t i) except *:
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex.at(i), LexDist_upper)
 | 
			
		||||
        bounds_check(i, self.length, PADDING)
 | 
			
		||||
        return lexeme_check_dist_flag(self.lex[i], LexDist_upper)
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in New Issue
	
	Block a user