Fix test after updating the French tokenizer stuff

This commit is contained in:
Matthew Honnibal 2017-02-27 11:20:47 +01:00
parent 26446aa728
commit 0aaa546435

View File

@ -3,7 +3,7 @@ from __future__ import unicode_literals
from ...language import Language
from ...attrs import LANG
from ...fr.language_data import TOKENIZER_EXCEPTIONS, STOP_WORDS
from ...fr.language_data import get_tokenizer_exceptions, STOP_WORDS
from ...language_data.punctuation import TOKENIZER_INFIXES, ALPHA
import pytest
@ -20,7 +20,7 @@ def fr_tokenizer_w_infix():
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'fr'
tokenizer_exceptions = TOKENIZER_EXCEPTIONS
tokenizer_exceptions = get_tokenizer_exceptions()
stop_words = STOP_WORDS
infixes = TOKENIZER_INFIXES + [SPLIT_INFIX]