spaCy/spacy/tests/lang/fi/test_tokenizer.py

18 lines
564 B
Python
Raw Normal View History

2017-03-12 15:07:28 +03:00
# coding: utf8
from __future__ import unicode_literals
import pytest
ABBREVIATION_TESTS = [
('Hyvää uutta vuotta t. siht. Niemelä!', ['Hyvää', 'uutta', 'vuotta', 't.', 'siht.', 'Niemelä', '!']),
('Paino on n. 2.2 kg', ['Paino', 'on', 'n.', '2.2', 'kg'])
]
@pytest.mark.parametrize('text,expected_tokens', ABBREVIATION_TESTS)
def test_tokenizer_handles_testcases(fi_tokenizer, text, expected_tokens):
tokens = fi_tokenizer(text)
token_list = [token.text for token in tokens if not token.is_space]
assert expected_tokens == token_list