diff --git a/spacy/lang/tokenizer_exceptions.py b/spacy/lang/tokenizer_exceptions.py index f9b6530ec..6a7c09a44 100644 --- a/spacy/lang/tokenizer_exceptions.py +++ b/spacy/lang/tokenizer_exceptions.py @@ -3,7 +3,7 @@ from __future__ import unicode_literals # The use of this module turns out to be important, to avoid pathological # back-tracking. See Issue #957 -import regex +import regex as re from ..symbols import ORTH, POS, LEMMA, SPACE, PUNCT @@ -53,7 +53,7 @@ URL_PATTERN = ( r"$" ).strip() -TOKEN_MATCH = regex.compile(URL_PATTERN, regex.UNICODE).match +TOKEN_MATCH = re.compile(URL_PATTERN, re.UNICODE).match diff --git a/spacy/orth.pyx b/spacy/orth.pyx index a34058b8e..da5d95735 100644 --- a/spacy/orth.pyx +++ b/spacy/orth.pyx @@ -3,7 +3,7 @@ from __future__ import unicode_literals import unicodedata -import re +import regex as re # Binary string features