From d17e7dca9ee36a33c6d23bc8786b3c76ca9dc061 Mon Sep 17 00:00:00 2001 From: Ines Montani Date: Sat, 21 Dec 2019 19:57:41 +0100 Subject: [PATCH] Fix problems caused by merge conflict --- spacy/scorer.py | 2 +- spacy/tokenizer.pyx | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/spacy/scorer.py b/spacy/scorer.py index 7fee4865a..6238b6ead 100644 --- a/spacy/scorer.py +++ b/spacy/scorer.py @@ -313,7 +313,7 @@ class Scorer(object): cand_deps_per_dep[token.dep_.lower()].add( (gold_i, gold_head, token.dep_.lower()) ) - if "-" not in [token[-1] for token in gold.orig_annot]: + if "-" not in [token[-1] for token in orig.entities]: # Find all NER labels in gold and doc ent_labels = set([x[0] for x in gold_ents] + [k.label_ for k in doc.ents]) # Set up all labels for per type scoring and prepare gold per type diff --git a/spacy/tokenizer.pyx b/spacy/tokenizer.pyx index c1ac3dd06..f0120c708 100644 --- a/spacy/tokenizer.pyx +++ b/spacy/tokenizer.pyx @@ -116,10 +116,10 @@ cdef class Tokenizer: def __set__(self, rules): self._rules = {} self._reset_cache([key for key in self._cache]) - self._reset_specials() + self._flush_specials() self._cache = PreshMap() self._specials = PreshMap() - self._load_special_tokenization(rules) + self._load_special_cases(rules) def __reduce__(self): args = (self.vocab,