2017-02-04 14:47:29 +03:00
|
|
|
import pytest
|
|
|
|
|
2017-05-09 01:02:37 +03:00
|
|
|
|
2017-02-04 14:47:29 +03:00
|
|
|
ABBREVIATION_TESTS = [
|
2018-11-27 03:09:36 +03:00
|
|
|
(
|
|
|
|
"Hyvää uutta vuotta t. siht. Niemelä!",
|
|
|
|
["Hyvää", "uutta", "vuotta", "t.", "siht.", "Niemelä", "!"],
|
|
|
|
),
|
|
|
|
("Paino on n. 2.2 kg", ["Paino", "on", "n.", "2.2", "kg"]),
|
2020-02-22 16:12:32 +03:00
|
|
|
(
|
|
|
|
"Vuonna 1 eaa. tapahtui kauheita.",
|
|
|
|
["Vuonna", "1", "eaa.", "tapahtui", "kauheita", "."],
|
|
|
|
),
|
2017-02-04 14:47:29 +03:00
|
|
|
]
|
|
|
|
|
2019-12-03 14:55:28 +03:00
|
|
|
HYPHENATED_TESTS = [
|
|
|
|
(
|
2020-02-22 16:12:32 +03:00
|
|
|
"1700-luvulle sijoittuva taide-elokuva Wikimedia-säätiön Varsinais-Suomen",
|
|
|
|
[
|
|
|
|
"1700-luvulle",
|
|
|
|
"sijoittuva",
|
|
|
|
"taide-elokuva",
|
|
|
|
"Wikimedia-säätiön",
|
|
|
|
"Varsinais-Suomen",
|
|
|
|
],
|
2019-12-03 14:55:28 +03:00
|
|
|
)
|
|
|
|
]
|
|
|
|
|
2020-02-11 04:32:43 +03:00
|
|
|
ABBREVIATION_INFLECTION_TESTS = [
|
|
|
|
(
|
|
|
|
"VTT:ssa ennen v:ta 2010 suoritetut mittaukset",
|
2020-02-18 17:38:18 +03:00
|
|
|
["VTT:ssa", "ennen", "v:ta", "2010", "suoritetut", "mittaukset"],
|
2020-02-11 04:32:43 +03:00
|
|
|
),
|
2020-02-18 17:38:18 +03:00
|
|
|
("ALV:n osuus on 24 %.", ["ALV:n", "osuus", "on", "24", "%", "."]),
|
|
|
|
("Hiihtäjä oli kilpailun 14:s.", ["Hiihtäjä", "oli", "kilpailun", "14:s", "."]),
|
2020-02-22 16:12:32 +03:00
|
|
|
("EU:n toimesta tehtiin jotain.", ["EU:n", "toimesta", "tehtiin", "jotain", "."]),
|
2020-02-11 04:32:43 +03:00
|
|
|
]
|
|
|
|
|
2021-06-16 11:56:47 +03:00
|
|
|
CONTRACTION_TESTS = [
|
|
|
|
(
|
|
|
|
"Päätimme ettemme tule.",
|
|
|
|
["Päätimme", "ett", "emme", "tule", "."],
|
|
|
|
["päätimme", "että", "emme", "tule", "."]
|
|
|
|
),
|
|
|
|
(
|
|
|
|
"Miksei puhuttaisi?",
|
|
|
|
["Miks", "ei", "puhuttaisi", "?"],
|
|
|
|
["miksi", "ei", "puhuttaisi", "?"]
|
|
|
|
),
|
|
|
|
(
|
|
|
|
"He tottelivat vaikkeivat halunneet",
|
|
|
|
["He", "tottelivat", "vaikk", "eivat", "halunneet"],
|
|
|
|
["he", "tottelivat", "vaikka", "eivät", "halunneet"]
|
|
|
|
),
|
|
|
|
]
|
2017-02-04 14:47:29 +03:00
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("text,expected_tokens", ABBREVIATION_TESTS)
|
2019-12-03 14:55:28 +03:00
|
|
|
def test_fi_tokenizer_abbreviations(fi_tokenizer, text, expected_tokens):
|
|
|
|
tokens = fi_tokenizer(text)
|
|
|
|
token_list = [token.text for token in tokens if not token.is_space]
|
|
|
|
assert expected_tokens == token_list
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize("text,expected_tokens", HYPHENATED_TESTS)
|
|
|
|
def test_fi_tokenizer_hyphenated_words(fi_tokenizer, text, expected_tokens):
|
2017-02-04 14:47:29 +03:00
|
|
|
tokens = fi_tokenizer(text)
|
|
|
|
token_list = [token.text for token in tokens if not token.is_space]
|
|
|
|
assert expected_tokens == token_list
|
2020-02-11 04:32:43 +03:00
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize("text,expected_tokens", ABBREVIATION_INFLECTION_TESTS)
|
|
|
|
def test_fi_tokenizer_abbreviation_inflections(fi_tokenizer, text, expected_tokens):
|
|
|
|
tokens = fi_tokenizer(text)
|
|
|
|
token_list = [token.text for token in tokens if not token.is_space]
|
|
|
|
assert expected_tokens == token_list
|
2021-06-16 11:56:47 +03:00
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.parametrize("text,expected_tokens,expected_norms", CONTRACTION_TESTS)
|
|
|
|
def test_fi_tokenizer_contractions(fi_tokenizer, text, expected_tokens, expected_norms):
|
|
|
|
tokens = fi_tokenizer(text)
|
|
|
|
token_list = [token.text for token in tokens if not token.is_space]
|
|
|
|
norm_list = [token.norm_ for token in tokens if not token.is_space]
|
|
|
|
assert expected_tokens == token_list
|
|
|
|
assert expected_norms == norm_list
|