mirror of
https://github.com/explosion/spaCy.git
synced 2025-06-13 01:23:04 +03:00
Mark lemmatizer tests as models since they use installed data
This commit is contained in:
parent
138deb80a1
commit
49186b34a1
|
@ -6,6 +6,7 @@ from ...lemmatizer import read_index, read_exc
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
@pytest.mark.parametrize('text,lemmas', [("aardwolves", ["aardwolf"]),
|
@pytest.mark.parametrize('text,lemmas', [("aardwolves", ["aardwolf"]),
|
||||||
("aardwolf", ["aardwolf"]),
|
("aardwolf", ["aardwolf"]),
|
||||||
("planets", ["planet"]),
|
("planets", ["planet"]),
|
||||||
|
@ -17,6 +18,7 @@ def test_tagger_lemmatizer_noun_lemmas(lemmatizer, text, lemmas):
|
||||||
assert lemmatizer.noun(text) == set(lemmas)
|
assert lemmatizer.noun(text) == set(lemmas)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
def test_tagger_lemmatizer_base_forms(lemmatizer):
|
def test_tagger_lemmatizer_base_forms(lemmatizer):
|
||||||
if lemmatizer is None:
|
if lemmatizer is None:
|
||||||
return None
|
return None
|
||||||
|
@ -24,12 +26,14 @@ def test_tagger_lemmatizer_base_forms(lemmatizer):
|
||||||
assert lemmatizer.noun('dive', {'number': 'plur'}) == set(['diva'])
|
assert lemmatizer.noun('dive', {'number': 'plur'}) == set(['diva'])
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
def test_tagger_lemmatizer_base_form_verb(lemmatizer):
|
def test_tagger_lemmatizer_base_form_verb(lemmatizer):
|
||||||
if lemmatizer is None:
|
if lemmatizer is None:
|
||||||
return None
|
return None
|
||||||
assert lemmatizer.verb('saw', {'verbform': 'past'}) == set(['see'])
|
assert lemmatizer.verb('saw', {'verbform': 'past'}) == set(['see'])
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
def test_tagger_lemmatizer_punct(lemmatizer):
|
def test_tagger_lemmatizer_punct(lemmatizer):
|
||||||
if lemmatizer is None:
|
if lemmatizer is None:
|
||||||
return None
|
return None
|
||||||
|
@ -37,6 +41,7 @@ def test_tagger_lemmatizer_punct(lemmatizer):
|
||||||
assert lemmatizer.punct('“') == set(['"'])
|
assert lemmatizer.punct('“') == set(['"'])
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
def test_tagger_lemmatizer_read_index(path):
|
def test_tagger_lemmatizer_read_index(path):
|
||||||
if path is not None:
|
if path is not None:
|
||||||
with (path / 'wordnet' / 'index.noun').open() as file_:
|
with (path / 'wordnet' / 'index.noun').open() as file_:
|
||||||
|
@ -46,6 +51,7 @@ def test_tagger_lemmatizer_read_index(path):
|
||||||
assert 'plant' in index
|
assert 'plant' in index
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.models
|
||||||
@pytest.mark.parametrize('text,lemma', [("was", "be")])
|
@pytest.mark.parametrize('text,lemma', [("was", "be")])
|
||||||
def test_tagger_lemmatizer_read_exc(path, text, lemma):
|
def test_tagger_lemmatizer_read_exc(path, text, lemma):
|
||||||
if path is not None:
|
if path is not None:
|
||||||
|
|
Loading…
Reference in New Issue
Block a user