diff --git a/spacy/tests/conftest.py b/spacy/tests/conftest.py index c339f66ef..c5ea99007 100644 --- a/spacy/tests/conftest.py +++ b/spacy/tests/conftest.py @@ -1,14 +1,64 @@ -import pytest -import os +# coding: utf-8 +from __future__ import unicode_literals from ..en import English from ..de import German +from ..es import Spanish +from ..it import Italian +from ..fr import French +from ..pt import Portuguese +from ..nl import Dutch +from ..sv import Swedish +from ..hu import Hungarian +from ..tokens import Doc +from ..attrs import ORTH, TAG, HEAD, DEP + +from StringIO import StringIO +import pytest +LANGUAGES = [English, German, Spanish, Italian, French, Portuguese, Dutch, + Swedish, Hungarian] + + +@pytest.fixture(params=LANGUAGES) +def tokenizer(request): + lang = request.param + return lang.Defaults.create_tokenizer() + + +@pytest.fixture +def en_tokenizer(): + return English.Defaults.create_tokenizer() + + +@pytest.fixture +def en_vocab(): + return English.Defaults.create_vocab() + + +@pytest.fixture +def de_tokenizer(): + return German.Defaults.create_tokenizer() + + +@pytest.fixture +def hu_tokenizer(): + return Hungarian.Defaults.create_tokenizer() + + +@pytest.fixture +def text_file(): + return StringIO() + + +# deprecated, to be replaced with more specific instances @pytest.fixture(scope="session") def EN(): return English() + +# deprecated, to be replaced with more specific instances @pytest.fixture(scope="session") def DE(): return German() diff --git a/spacy/tests/de/conftest.py b/spacy/tests/de/conftest.py deleted file mode 100644 index c6b8be26e..000000000 --- a/spacy/tests/de/conftest.py +++ /dev/null @@ -1,11 +0,0 @@ -# coding: utf-8 -from __future__ import unicode_literals - -import pytest - -from ...de import German - - -@pytest.fixture -def de_tokenizer(): - return German.Defaults.create_tokenizer() diff --git a/spacy/tests/en/conftest.py b/spacy/tests/en/conftest.py deleted file mode 100644 index 3a3516c41..000000000 --- a/spacy/tests/en/conftest.py +++ /dev/null @@ -1,11 +0,0 @@ -# coding: utf-8 -from __future__ import unicode_literals - -import pytest - -from ...en import English - - -@pytest.fixture -def en_tokenizer(): - return English.Defaults.create_tokenizer() diff --git a/spacy/tests/hu/conftest.py b/spacy/tests/hu/conftest.py deleted file mode 100644 index 222bd1b00..000000000 --- a/spacy/tests/hu/conftest.py +++ /dev/null @@ -1,11 +0,0 @@ -# coding: utf-8 -from __future__ import unicode_literals - -import pytest - -from ...hu import Hungarian - - -@pytest.fixture -def hu_tokenizer(): - return Hungarian.Defaults.create_tokenizer() diff --git a/spacy/tests/tokenizer/conftest.py b/spacy/tests/tokenizer/conftest.py deleted file mode 100644 index c8e340208..000000000 --- a/spacy/tests/tokenizer/conftest.py +++ /dev/null @@ -1,23 +0,0 @@ -# coding: utf-8 -from __future__ import unicode_literals - -import pytest - -from ...en import English -from ...de import German -from ...es import Spanish -from ...it import Italian -from ...fr import French -from ...pt import Portuguese -from ...nl import Dutch -from ...sv import Swedish -from ...hu import Hungarian - - -LANGUAGES = [English, German, Spanish, Italian, French, Dutch, Swedish, Hungarian] - - -@pytest.fixture(params=LANGUAGES) -def tokenizer(request): - lang = request.param - return lang.Defaults.create_tokenizer()