From 34ca59691b6db947679552e899ca7209e41451db Mon Sep 17 00:00:00 2001 From: Jim O'Regan Date: Tue, 31 Oct 2017 14:50:13 +0000 Subject: [PATCH] no idea what is wrong here --- spacy/tests/conftest.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/spacy/tests/conftest.py b/spacy/tests/conftest.py index 5fa0c0cb7..ee4093db3 100644 --- a/spacy/tests/conftest.py +++ b/spacy/tests/conftest.py @@ -14,9 +14,8 @@ from .. import util # These languages are used for generic tokenizer tests – only add a language # here if it's using spaCy's tokenizer (not a different library) # TODO: re-implement generic tokenizer tests -_languages = ['bn', 'da', 'de', 'en', 'es', 'fi', 'fr', 'he', 'hu', 'id', - 'it', 'nb', 'nl', 'pl', 'pt', 'sv', 'ga', 'xx'] - +_languages = ['bn', 'da', 'de', 'en', 'es', 'fi', 'fr', 'ga', 'he', 'hu', 'id', + 'it', 'nb', 'nl', 'pl', 'pt', 'sv', 'xx'] _models = {'en': ['en_core_web_sm'], 'de': ['de_core_news_md'], 'fr': ['fr_depvec_web_lg'], @@ -107,6 +106,7 @@ def sv_tokenizer(): def bn_tokenizer(): return util.get_lang_class('bn').Defaults.create_tokenizer() + @pytest.fixture def ga_tokenizer(): return util.get_lang_class('ga').Defaults.create_tokenizer()