From e7ef51b3828f9acf0dc815c2d5eeddca5eda3d28 Mon Sep 17 00:00:00 2001 From: Ines Montani Date: Fri, 2 Jun 2017 19:00:01 +0200 Subject: [PATCH] Update tokenizer_exceptions.py --- spacy/es/tokenizer_exceptions.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/spacy/es/tokenizer_exceptions.py b/spacy/es/tokenizer_exceptions.py index fb274f907..f9c955338 100644 --- a/spacy/es/tokenizer_exceptions.py +++ b/spacy/es/tokenizer_exceptions.py @@ -6,10 +6,9 @@ from ..language_data import PRON_LEMMA, DET_LEMMA TOKENIZER_EXCEPTIONS = { - "pal": [ {ORTH: "pa", LEMMA: "para"}, - {ORTH: "el", LEMMA: DET_LEMMA, NORM: "el"} + {ORTH: "l", LEMMA: DET_LEMMA, NORM: "el"} ], "pala": [