From 84a87951eb8f97eff9992a3fa79a8ae5c8a6e2f3 Mon Sep 17 00:00:00 2001 From: Ines Montani Date: Tue, 3 Jan 2017 18:27:43 +0100 Subject: [PATCH] Fix typos --- spacy/en/tokenizer_exceptions.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/spacy/en/tokenizer_exceptions.py b/spacy/en/tokenizer_exceptions.py index 4df3fe535..9537761b3 100644 --- a/spacy/en/tokenizer_exceptions.py +++ b/spacy/en/tokenizer_exceptions.py @@ -19,18 +19,18 @@ for pron in ["i"]: {ORTH: "'m", LEMMA: "be", TAG: "VBP", "tenspect": 1, "number": 1} ] - EXC[pron + "m"] = [ + EXC[orth + "m"] = [ {ORTH: pron, LEMMA: PRON_LEMMA, TAG: "PRP"}, {ORTH: "m", LEMMA: "be", TAG: "VBP", "tenspect": 1, "number": 1 } ] - EXC[pron + "'ma"] = [ + EXC[orth + "'ma"] = [ {ORTH: pron, LEMMA: PRON_LEMMA, TAG: "PRP"}, {ORTH: "'m", LEMMA: "be", NORM: "am"}, {ORTH: "a", LEMMA: "going to", NORM: "gonna"} ] - EXC[pron + "ma"] = [ + EXC[orth + "ma"] = [ {ORTH: pron, LEMMA: PRON_LEMMA, TAG: "PRP"}, {ORTH: "m", LEMMA: "be", NORM: "am"}, {ORTH: "a", LEMMA: "going to", NORM: "gonna"}