Fix customized-tokenizer tests

This commit is contained in:
Matthew Honnibal 2017-09-04 20:13:13 +02:00
parent 34c585396a
commit 45029a550e

View File

@ -9,8 +9,8 @@ import pytest
@pytest.fixture @pytest.fixture
def tokenizer(en_vocab): def tokenizer(en_vocab):
prefix_re = util.compile_prefix_regex(nlp_model.Defaults.prefixes) prefix_re = util.compile_prefix_regex(English.Defaults.prefixes)
suffix_re = util.compile_suffix_regex(nlp_model.Defaults.suffixes) suffix_re = util.compile_suffix_regex(English.Defaults.suffixes)
custom_infixes = ['\.\.\.+', custom_infixes = ['\.\.\.+',
'(?<=[0-9])-(?=[0-9])', '(?<=[0-9])-(?=[0-9])',
# '(?<=[0-9]+),(?=[0-9]+)', # '(?<=[0-9]+),(?=[0-9]+)',