mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-14 13:47:13 +03:00
ec62cadf4c
* Add back Romanian in conftest * Romanian lex_attr * More tokenizer exceptions for Romanian * Add tests for some Romanian tokenizer exceptions
26 lines
806 B
Python
26 lines
806 B
Python
# coding: utf8
|
|
from __future__ import unicode_literals
|
|
|
|
import pytest
|
|
|
|
DEFAULT_TESTS = [
|
|
('Adresa este str. Principală nr. 5.', ['Adresa', 'este', 'str.', 'Principală', 'nr.', '5', '.']),
|
|
('Teste, etc.', ['Teste', ',', 'etc.']),
|
|
('Lista, ș.a.m.d.', ['Lista', ',', 'ș.a.m.d.']),
|
|
('Și d.p.d.v. al...', ['Și', 'd.p.d.v.', 'al', '...'])
|
|
]
|
|
|
|
NUMBER_TESTS = [
|
|
('Clasa a 4-a.', ['Clasa', 'a', '4-a', '.']),
|
|
('Al 12-lea ceas.', ['Al', '12-lea', 'ceas', '.'])
|
|
]
|
|
|
|
TESTCASES = DEFAULT_TESTS + NUMBER_TESTS
|
|
|
|
|
|
@pytest.mark.parametrize('text,expected_tokens', TESTCASES)
|
|
def test_tokenizer_handles_testcases(ro_tokenizer, text, expected_tokens):
|
|
tokens = ro_tokenizer(text)
|
|
token_list = [token.text for token in tokens if not token.is_space]
|
|
assert expected_tokens == token_list
|