diff --git a/spacy/tests/lang/ja/test_tokenizer.py b/spacy/tests/lang/ja/test_tokenizer.py index 7b322293a..e61975551 100644 --- a/spacy/tests/lang/ja/test_tokenizer.py +++ b/spacy/tests/lang/ja/test_tokenizer.py @@ -130,7 +130,7 @@ def test_ja_tokenizer_sub_tokens( [ ( "取ってつけた", - ("五段-ラ行,連用形-促音便", "", "下一段-カ行,連用形-一般", "助動詞-タ,終止形-一般"), + ("五段-ラ行;連用形-促音便", "", "下一段-カ行;連用形-一般", "助動詞-タ;終止形-一般"), ("トッ", "テ", "ツケ", "タ"), ), ], @@ -139,7 +139,7 @@ def test_ja_tokenizer_inflections_reading_forms( ja_tokenizer, text, inflections, reading_forms ): tokens = ja_tokenizer(text) - test_inflections = [",".join(tt.morph.get("inflection")) for tt in tokens] + test_inflections = [tt.morph.get("inflection")[0] for tt in tokens] assert test_inflections == list(inflections) test_readings = [tt.morph.get("reading")[0] for tt in tokens] assert test_readings == list(reading_forms)