mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-11 04:08:09 +03:00
renamming tests
This commit is contained in:
parent
c085c2d391
commit
26e31afc18
|
@ -4,15 +4,17 @@ from __future__ import unicode_literals
|
|||
|
||||
import pytest
|
||||
|
||||
|
||||
@pytest.mark.xfail
|
||||
def test_lemmatizer_(fr_tokenizer):
|
||||
def test_lemmatizer_verb(fr_tokenizer):
|
||||
text = "Je suis allé au mois de janv. aux prud’hommes."
|
||||
tokens = fr_tokenizer(text)
|
||||
assert len(tokens) == 10
|
||||
assert tokens[2].lemma_ == "aller"
|
||||
|
||||
|
||||
@pytest.mark.xfail
|
||||
def test_tokenizer_handles_exc_in_text_2(fr_tokenizer):
|
||||
def test_tokenizer_verb_2(fr_tokenizer):
|
||||
text = "Je dois manger ce soir"
|
||||
tokens = fr_tokenizer(text)
|
||||
assert len(tokens) == 11
|
||||
|
@ -20,14 +22,7 @@ def test_tokenizer_handles_exc_in_text_2(fr_tokenizer):
|
|||
|
||||
|
||||
@pytest.mark.xfail
|
||||
def test_tokenizer_handles_exc_in_text_2(fr_tokenizer):
|
||||
text = "Je dois manger ce soir"
|
||||
tokens = fr_tokenizer(text)
|
||||
assert len(tokens) == 11
|
||||
assert tokens[1].lemma_ == "devoir"
|
||||
|
||||
@pytest.mark.xfail
|
||||
def test_tokenizer_handles_exc_in_text_2(fr_tokenizer):
|
||||
def test_tokenizer_verb_noun(fr_tokenizer):
|
||||
# This one is tricky because notes is a NOUN and can be a VERB
|
||||
text = "Nous validerons vos notes plus tard"
|
||||
tokens = fr_tokenizer(text)
|
||||
|
|
Loading…
Reference in New Issue
Block a user