mirror of
https://github.com/explosion/spaCy.git
synced 2025-07-04 11:53:09 +03:00
fix: tokenizer.pxd
This commit is contained in:
parent
c2f0256606
commit
25ce9f48c6
|
@ -31,24 +31,58 @@ cdef class Tokenizer:
|
||||||
|
|
||||||
cdef Doc _tokenize_affixes(self, str string, bint with_special_cases)
|
cdef Doc _tokenize_affixes(self, str string, bint with_special_cases)
|
||||||
cdef int _apply_special_cases(self, Doc doc) except -1
|
cdef int _apply_special_cases(self, Doc doc) except -1
|
||||||
cdef void _filter_special_spans(self, vector[SpanC] &original,
|
cdef void _filter_special_spans(
|
||||||
vector[SpanC] &filtered, int doc_len) nogil
|
self,
|
||||||
cdef object _prepare_special_spans(self, Doc doc,
|
vector[SpanC] &original,
|
||||||
vector[SpanC] &filtered)
|
vector[SpanC] &filtered,
|
||||||
cdef int _retokenize_special_spans(self, Doc doc, TokenC* tokens,
|
int doc_len,
|
||||||
object span_data)
|
) nogil
|
||||||
cdef int _try_specials_and_cache(self, hash_t key, Doc tokens,
|
cdef object _prepare_special_spans(
|
||||||
|
self,
|
||||||
|
Doc doc,
|
||||||
|
vector[SpanC] &filtered,
|
||||||
|
)
|
||||||
|
cdef int _retokenize_special_spans(
|
||||||
|
self,
|
||||||
|
Doc doc,
|
||||||
|
TokenC* tokens,
|
||||||
|
object span_data,
|
||||||
|
)
|
||||||
|
cdef int _try_specials_and_cache(
|
||||||
|
self,
|
||||||
|
hash_t key,
|
||||||
|
Doc tokens,
|
||||||
int* has_special,
|
int* has_special,
|
||||||
bint with_special_cases) except -1
|
bint with_special_cases,
|
||||||
cdef int _tokenize(self, Doc tokens, str span, hash_t key,
|
) except -1
|
||||||
int* has_special, bint with_special_cases) except -1
|
cdef int _tokenize(
|
||||||
cdef str _split_affixes(self, Pool mem, str string,
|
self,
|
||||||
|
Doc tokens,
|
||||||
|
str span,
|
||||||
|
hash_t key,
|
||||||
|
int* has_special,
|
||||||
|
bint with_special_cases,
|
||||||
|
) except -1
|
||||||
|
cdef str _split_affixes(
|
||||||
|
self,
|
||||||
|
Pool mem,
|
||||||
|
str string,
|
||||||
vector[LexemeC*] *prefixes,
|
vector[LexemeC*] *prefixes,
|
||||||
vector[LexemeC*] *suffixes, int* has_special,
|
vector[LexemeC*] *suffixes, int* has_special,
|
||||||
bint with_special_cases)
|
bint with_special_cases,
|
||||||
cdef int _attach_tokens(self, Doc tokens, str string,
|
)
|
||||||
|
cdef int _attach_tokens(
|
||||||
|
self,
|
||||||
|
Doc tokens,
|
||||||
|
str string,
|
||||||
vector[LexemeC*] *prefixes,
|
vector[LexemeC*] *prefixes,
|
||||||
vector[LexemeC*] *suffixes, int* has_special,
|
vector[LexemeC*] *suffixes, int* has_special,
|
||||||
bint with_special_cases) except -1
|
bint with_special_cases,
|
||||||
cdef int _save_cached(self, const TokenC* tokens, hash_t key,
|
) except -1
|
||||||
int* has_special, int n) except -1
|
cdef int _save_cached(
|
||||||
|
self,
|
||||||
|
const TokenC* tokens,
|
||||||
|
hash_t key,
|
||||||
|
int* has_special,
|
||||||
|
int n,
|
||||||
|
) except -1
|
||||||
|
|
Loading…
Reference in New Issue
Block a user