mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-27 18:36:36 +03:00
ece30c28a8
This way, the tokenizer matches the tokenization in German treebanks
21 lines
701 B
Python
21 lines
701 B
Python
# coding: utf8
|
|
from __future__ import unicode_literals
|
|
|
|
from ..char_classes import LIST_ELLIPSES, LIST_ICONS
|
|
from ..char_classes import QUOTES, ALPHA, ALPHA_LOWER, ALPHA_UPPER
|
|
|
|
|
|
_quotes = QUOTES.replace("'", '')
|
|
|
|
_infixes = (LIST_ELLIPSES + LIST_ICONS +
|
|
[r'(?<=[{}])\.(?=[{}])'.format(ALPHA_LOWER, ALPHA_UPPER),
|
|
r'(?<=[{a}])[,!?](?=[{a}])'.format(a=ALPHA),
|
|
r'(?<=[{a}"])[:<>=](?=[{a}])'.format(a=ALPHA),
|
|
r'(?<=[{a}]),(?=[{a}])'.format(a=ALPHA),
|
|
r'(?<=[{a}])([{q}\)\]\(\[])(?=[\{a}])'.format(a=ALPHA, q=_quotes),
|
|
r'(?<=[{a}])--(?=[{a}])'.format(a=ALPHA),
|
|
r'(?<=[0-9])-(?=[0-9])'])
|
|
|
|
|
|
TOKENIZER_INFIXES = _infixes
|