From acb9bdb852da4225e5a58b9896c4f079b1038bd0 Mon Sep 17 00:00:00 2001 From: ines Date: Mon, 6 Nov 2017 17:41:53 +0100 Subject: [PATCH] Fix PRON_LEMMA imports --- spacy/deprecated.py | 4 ---- spacy/lang/bn/morph_rules.py | 3 +-- spacy/lang/da/morph_rules.py | 4 ++-- spacy/lang/de/tokenizer_exceptions.py | 3 +-- spacy/lang/en/morph_rules.py | 3 +-- spacy/lang/en/tokenizer_exceptions.py | 3 +-- spacy/lang/es/tokenizer_exceptions.py | 3 +-- spacy/lang/fr/tokenizer_exceptions.py | 3 +-- spacy/lang/nb/morph_rules.py | 3 +-- spacy/lang/pt/tokenizer_exceptions.py | 3 +-- spacy/lang/sv/morph_rules.py | 3 +-- spacy/lang/sv/tokenizer_exceptions.py | 3 +-- 12 files changed, 12 insertions(+), 26 deletions(-) delete mode 100644 spacy/deprecated.py diff --git a/spacy/deprecated.py b/spacy/deprecated.py deleted file mode 100644 index eafdbe580..000000000 --- a/spacy/deprecated.py +++ /dev/null @@ -1,4 +0,0 @@ -# coding: utf8 -from __future__ import unicode_literals - -PRON_LEMMA = "-PRON-" diff --git a/spacy/lang/bn/morph_rules.py b/spacy/lang/bn/morph_rules.py index 6ca8fc097..9e5511344 100644 --- a/spacy/lang/bn/morph_rules.py +++ b/spacy/lang/bn/morph_rules.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...deprecated import PRON_LEMMA -from ...symbols import LEMMA +from ...symbols import LEMMA, PRON_LEMMA MORPH_RULES = { diff --git a/spacy/lang/da/morph_rules.py b/spacy/lang/da/morph_rules.py index b365bf871..69ab3e404 100644 --- a/spacy/lang/da/morph_rules.py +++ b/spacy/lang/da/morph_rules.py @@ -1,8 +1,8 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import LEMMA -from ...deprecated import PRON_LEMMA +from ...symbols import LEMMA, PRON_LEMMA + MORPH_RULES = { "PRON": { diff --git a/spacy/lang/de/tokenizer_exceptions.py b/spacy/lang/de/tokenizer_exceptions.py index cb16fb06c..8e041a740 100644 --- a/spacy/lang/de/tokenizer_exceptions.py +++ b/spacy/lang/de/tokenizer_exceptions.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import ORTH, LEMMA, TAG, NORM -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, TAG, NORM, PRON_LEMMA _exc = { diff --git a/spacy/lang/en/morph_rules.py b/spacy/lang/en/morph_rules.py index fedb89700..002dc6805 100644 --- a/spacy/lang/en/morph_rules.py +++ b/spacy/lang/en/morph_rules.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import LEMMA -from ...deprecated import PRON_LEMMA +from ...symbols import LEMMA, PRON_LEMMA MORPH_RULES = { diff --git a/spacy/lang/en/tokenizer_exceptions.py b/spacy/lang/en/tokenizer_exceptions.py index a76b5fb2b..e91b5d9d1 100644 --- a/spacy/lang/en/tokenizer_exceptions.py +++ b/spacy/lang/en/tokenizer_exceptions.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import ORTH, LEMMA, TAG, NORM -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, TAG, NORM, PRON_LEMMA _exc = {} diff --git a/spacy/lang/es/tokenizer_exceptions.py b/spacy/lang/es/tokenizer_exceptions.py index d4131ddf6..488b94762 100644 --- a/spacy/lang/es/tokenizer_exceptions.py +++ b/spacy/lang/es/tokenizer_exceptions.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import ORTH, LEMMA, TAG, NORM, ADP, DET -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, TAG, NORM, ADP, DET, PRON_LEMMA _exc = { diff --git a/spacy/lang/fr/tokenizer_exceptions.py b/spacy/lang/fr/tokenizer_exceptions.py index 9994686ac..d2db1447c 100644 --- a/spacy/lang/fr/tokenizer_exceptions.py +++ b/spacy/lang/fr/tokenizer_exceptions.py @@ -7,8 +7,7 @@ from ._tokenizer_exceptions_list import FR_BASE_EXCEPTIONS from .punctuation import ELISION, HYPHENS from ..tokenizer_exceptions import URL_PATTERN from ..char_classes import ALPHA_LOWER -from ...symbols import ORTH, LEMMA, TAG, NORM -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, TAG, NORM, PRON_LEMMA def upper_first_letter(text): diff --git a/spacy/lang/nb/morph_rules.py b/spacy/lang/nb/morph_rules.py index 924c9beb5..d1bea9ebc 100644 --- a/spacy/lang/nb/morph_rules.py +++ b/spacy/lang/nb/morph_rules.py @@ -1,8 +1,7 @@ # encoding: utf8 from __future__ import unicode_literals -from ...symbols import LEMMA -from ...deprecated import PRON_LEMMA +from ...symbols import LEMMA, PRON_LEMMA # Used the table of pronouns at https://no.wiktionary.org/wiki/Tillegg:Pronomen_i_norsk diff --git a/spacy/lang/pt/tokenizer_exceptions.py b/spacy/lang/pt/tokenizer_exceptions.py index 6e8b8a24c..69578655c 100644 --- a/spacy/lang/pt/tokenizer_exceptions.py +++ b/spacy/lang/pt/tokenizer_exceptions.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import ORTH, LEMMA, NORM -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, NORM, PRON_LEMMA _exc = { diff --git a/spacy/lang/sv/morph_rules.py b/spacy/lang/sv/morph_rules.py index 2875eb3c8..e28322e98 100644 --- a/spacy/lang/sv/morph_rules.py +++ b/spacy/lang/sv/morph_rules.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import LEMMA -from ...deprecated import PRON_LEMMA +from ...symbols import LEMMA, PRON_LEMMA # Used the table of pronouns at https://sv.wiktionary.org/wiki/deras diff --git a/spacy/lang/sv/tokenizer_exceptions.py b/spacy/lang/sv/tokenizer_exceptions.py index 64aedf8af..4f84ef9b5 100644 --- a/spacy/lang/sv/tokenizer_exceptions.py +++ b/spacy/lang/sv/tokenizer_exceptions.py @@ -1,8 +1,7 @@ # coding: utf8 from __future__ import unicode_literals -from ...symbols import ORTH, LEMMA, TAG, NORM -from ...deprecated import PRON_LEMMA +from ...symbols import ORTH, LEMMA, TAG, NORM, PRON_LEMMA _exc = {}