mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-24 00:46:28 +03:00
Improve Catalan tokenization accuracy (#3225)
* small hyphen clean up for French * catalan infix similar to french
This commit is contained in:
parent
e00680a33a
commit
a3efa3e8d9
|
@ -12,6 +12,8 @@ from ...language import Language
|
|||
from ...attrs import LANG, NORM
|
||||
from ...util import update_exc, add_lookups
|
||||
|
||||
from .punctuation import TOKENIZER_INFIXES
|
||||
|
||||
|
||||
class CatalanDefaults(Language.Defaults):
|
||||
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
|
||||
|
@ -23,6 +25,7 @@ class CatalanDefaults(Language.Defaults):
|
|||
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS, TOKENIZER_EXCEPTIONS)
|
||||
stop_words = STOP_WORDS
|
||||
lemma_lookup = LOOKUP
|
||||
infixes = TOKENIZER_INFIXES
|
||||
|
||||
|
||||
class Catalan(Language):
|
||||
|
|
15
spacy/lang/ca/punctuation.py
Normal file
15
spacy/lang/ca/punctuation.py
Normal file
|
@ -0,0 +1,15 @@
|
|||
# coding: utf8
|
||||
from __future__ import unicode_literals
|
||||
|
||||
from ..punctuation import TOKENIZER_INFIXES
|
||||
from ..char_classes import ALPHA
|
||||
|
||||
|
||||
ELISION = " ' ’ ".strip().replace(" ", "").replace("\n", "")
|
||||
|
||||
|
||||
_infixes = TOKENIZER_INFIXES + [
|
||||
r"(?<=[{a}][{el}])(?=[{a}])".format(a=ALPHA, el=ELISION)
|
||||
]
|
||||
|
||||
TOKENIZER_INFIXES = _infixes
|
|
@ -317,8 +317,6 @@ _hyphen_prefix = [
|
|||
"Vaux",
|
||||
]
|
||||
|
||||
_other_hyphens = "".join([h for h in HYPHENS if h != "-"])
|
||||
|
||||
_regular_exp = [
|
||||
"^a[{hyphen}]sexualis[{al}]+$".format(hyphen=HYPHENS, al=ALPHA_LOWER),
|
||||
"^arginine[{hyphen}]méthyl[{al}]+$".format(hyphen=HYPHENS, al=ALPHA_LOWER),
|
||||
|
@ -375,10 +373,9 @@ _regular_exp = [
|
|||
]
|
||||
# catching cases like faux-vampire
|
||||
_regular_exp += [
|
||||
"^{prefix}[{hyphen}][{al}][{al}{elision}{other_hyphen}\-]*$".format(
|
||||
"^{prefix}[{hyphen}][{al}][{hyphen}{al}{elision}]*$".format(
|
||||
prefix=p,
|
||||
hyphen=HYPHENS,
|
||||
other_hyphen=_other_hyphens,
|
||||
hyphen=HYPHENS, # putting the - first in the [] range avoids having to use a backslash
|
||||
elision=ELISION,
|
||||
al=ALPHA_LOWER,
|
||||
)
|
||||
|
@ -388,8 +385,8 @@ _regular_exp += [
|
|||
# catching cases like entr'abat
|
||||
_elision_prefix = ["r?é?entr", "grande?s?", "r"]
|
||||
_regular_exp += [
|
||||
"^{prefix}[{elision}][{al}][{al}{elision}{hyphen}\-]*$".format(
|
||||
prefix=p, elision=ELISION, hyphen=_other_hyphens, al=ALPHA_LOWER
|
||||
"^{prefix}[{elision}][{al}][{hyphen}{al}{elision}]*$".format(
|
||||
prefix=p, elision=ELISION, hyphen=HYPHENS, al=ALPHA_LOWER
|
||||
)
|
||||
for p in _elision_prefix
|
||||
]
|
||||
|
|
13
spacy/tests/lang/ca/test_prefix_suffix_infix.py
Normal file
13
spacy/tests/lang/ca/test_prefix_suffix_infix.py
Normal file
|
@ -0,0 +1,13 @@
|
|||
# coding: utf-8
|
||||
from __future__ import unicode_literals
|
||||
|
||||
import pytest
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
"text,expected_tokens", [("d'un", ["d'", "un"]), ("s'ha", ["s'", "ha"])]
|
||||
)
|
||||
def test_contractions(ca_tokenizer, text, expected_tokens):
|
||||
""" Test that the contractions are split into two tokens"""
|
||||
tokens = ca_tokenizer(text)
|
||||
assert len(tokens) == 2
|
|
@ -18,7 +18,7 @@ def test_ca_tokenizer_handles_long_text(ca_tokenizer):
|
|||
una gerra de cervesa. Ens asseiem -fotògraf i periodista- en una terrassa buida."""
|
||||
|
||||
tokens = ca_tokenizer(text)
|
||||
assert len(tokens) == 136
|
||||
assert len(tokens) == 138
|
||||
|
||||
|
||||
@pytest.mark.parametrize('text,length', [
|
||||
|
|
Loading…
Reference in New Issue
Block a user