spaCy/spacy/lang/es/tokenizer_exceptions.py

83 lines
1.4 KiB
Python
Raw Normal View History

from ...symbols import NORM, ORTH
from ...util import update_exc
from ..tokenizer_exceptions import BASE_EXCEPTIONS
2016-12-18 18:54:19 +03:00
2017-05-08 16:48:04 +03:00
_exc = {
"pal": [{ORTH: "pa"}, {ORTH: "l", NORM: "el"}],
2017-05-08 16:48:04 +03:00
}
2017-05-08 16:48:04 +03:00
for exc_data in [
{ORTH: ""},
{ORTH: "°C"},
{ORTH: "aprox."},
{ORTH: "dna."},
{ORTH: "dpto."},
{ORTH: "ej."},
{ORTH: "esq."},
{ORTH: "pág."},
{ORTH: "p.ej."},
{ORTH: "Ud.", NORM: "usted"},
{ORTH: "Vd.", NORM: "usted"},
{ORTH: "Uds.", NORM: "ustedes"},
{ORTH: "Vds.", NORM: "ustedes"},
{ORTH: "vol.", NORM: "volúmen"},
]:
2017-11-02 01:02:45 +03:00
_exc[exc_data[ORTH]] = [exc_data]
2017-05-08 16:48:04 +03:00
# Times
_exc["12m."] = [{ORTH: "12"}, {ORTH: "m."}]
2017-05-08 16:48:04 +03:00
for h in range(1, 12 + 1):
for period in ["a.m.", "am"]:
_exc[f"{h}{period}"] = [{ORTH: f"{h}"}, {ORTH: period}]
2017-05-08 16:48:04 +03:00
for period in ["p.m.", "pm"]:
_exc[f"{h}{period}"] = [{ORTH: f"{h}"}, {ORTH: period}]
2017-05-08 16:48:04 +03:00
for orth in [
"a.C.",
"a.J.C.",
"d.C.",
"d.J.C.",
"apdo.",
"Av.",
"Avda.",
"Cía.",
"Dr.",
"Dra.",
"EE.UU.",
"Ee.Uu.",
2021-01-29 16:27:22 +03:00
"EE. UU.",
"Ee. Uu.",
"etc.",
"fig.",
"Gob.",
"Gral.",
"Ing.",
"J.C.",
"km/h",
"Lic.",
"m.n.",
"núm.",
"P.D.",
"Prof.",
"Profa.",
"q.e.p.d.",
"Q.E.P.D.",
"S.A.",
"S.L.",
"S.R.L.",
"s.s.s.",
"Sr.",
"Sra.",
"Srta.",
]:
2017-05-08 16:48:04 +03:00
_exc[orth] = [{ORTH: orth}]
2016-12-18 18:54:19 +03:00
TOKENIZER_EXCEPTIONS = update_exc(BASE_EXCEPTIONS, _exc)