mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-27 18:36:36 +03:00
25 lines
798 B
Python
25 lines
798 B
Python
|
# coding: utf-8
|
||
|
|
||
|
from __future__ import unicode_literals
|
||
|
|
||
|
import pytest
|
||
|
|
||
|
|
||
|
@pytest.mark.parametrize('text,lemma', [("aprox.", "aproximadamente"),
|
||
|
("esq.", "esquina"),
|
||
|
("pág.", "página"),
|
||
|
("p.ej.", "por ejemplo")
|
||
|
])
|
||
|
def test_tokenizer_handles_abbr(es_tokenizer, text, lemma):
|
||
|
tokens = es_tokenizer(text)
|
||
|
assert len(tokens) == 1
|
||
|
assert tokens[0].lemma_ == lemma
|
||
|
|
||
|
|
||
|
def test_tokenizer_handles_exc_in_text(es_tokenizer):
|
||
|
text = "Mariano Rajoy ha corrido aprox. medio kilómetro"
|
||
|
tokens = es_tokenizer(text)
|
||
|
assert len(tokens) == 7
|
||
|
assert tokens[4].text == "aprox."
|
||
|
assert tokens[4].lemma_ == "aproximadamente"
|