mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-15 14:17:58 +03:00
03fefa37e2
* Add overwrite settings for more components For pipeline components where it's relevant and not already implemented, add an explicit `overwrite` setting that controls whether `set_annotations` overwrites existing annotation. For the `morphologizer`, add an additional setting `extend`, which controls whether the existing features are preserved. * +overwrite, +extend: overwrite values of existing features, add any new features * +overwrite, -extend: overwrite completely, removing any existing features * -overwrite, +extend: keep values of existing features, add any new features * -overwrite, -extend: do not modify the existing value if set In all cases an unset value will be set by `set_annotations`. Preserve current overwrite defaults: * True: morphologizer, entity linker * False: tagger, sentencizer, senter * Add backwards compat overwrite settings * Put empty line back Removed by accident in last commit * Set backwards-compatible defaults in __init__ Because the `TrainablePipe` serialization methods update `cfg`, there's no straightforward way to detect whether models serialized with a previous version are missing the overwrite settings. It would be possible in the sentencizer due to its separate serialization methods, however to keep the changes parallel, this also sets the default in `__init__`. * Remove traces Co-authored-by: Paul O'Leary McCann <polm@dampfkraft.com>
200 lines
6.9 KiB
Python
200 lines
6.9 KiB
Python
import pytest
|
|
from numpy.testing import assert_equal
|
|
|
|
from spacy import util
|
|
from spacy.training import Example
|
|
from spacy.lang.en import English
|
|
from spacy.language import Language
|
|
from spacy.tests.util import make_tempdir
|
|
from spacy.morphology import Morphology
|
|
from spacy.attrs import MORPH
|
|
from spacy.tokens import Doc
|
|
|
|
|
|
def test_label_types():
|
|
nlp = Language()
|
|
morphologizer = nlp.add_pipe("morphologizer")
|
|
morphologizer.add_label("Feat=A")
|
|
with pytest.raises(ValueError):
|
|
morphologizer.add_label(9)
|
|
|
|
|
|
TRAIN_DATA = [
|
|
(
|
|
"I like green eggs",
|
|
{
|
|
"morphs": ["Feat=N", "Feat=V", "Feat=J", "Feat=N"],
|
|
"pos": ["NOUN", "VERB", "ADJ", "NOUN"],
|
|
},
|
|
),
|
|
# test combinations of morph+POS
|
|
("Eat blue ham", {"morphs": ["Feat=V", "", ""], "pos": ["", "ADJ", ""]}),
|
|
]
|
|
|
|
|
|
def test_no_label():
|
|
nlp = Language()
|
|
nlp.add_pipe("morphologizer")
|
|
with pytest.raises(ValueError):
|
|
nlp.initialize()
|
|
|
|
|
|
def test_implicit_label():
|
|
nlp = Language()
|
|
nlp.add_pipe("morphologizer")
|
|
train_examples = []
|
|
for t in TRAIN_DATA:
|
|
train_examples.append(Example.from_dict(nlp.make_doc(t[0]), t[1]))
|
|
nlp.initialize(get_examples=lambda: train_examples)
|
|
|
|
|
|
def test_no_resize():
|
|
nlp = Language()
|
|
morphologizer = nlp.add_pipe("morphologizer")
|
|
morphologizer.add_label("POS" + Morphology.FIELD_SEP + "NOUN")
|
|
morphologizer.add_label("POS" + Morphology.FIELD_SEP + "VERB")
|
|
nlp.initialize()
|
|
# this throws an error because the morphologizer can't be resized after initialization
|
|
with pytest.raises(ValueError):
|
|
morphologizer.add_label("POS" + Morphology.FIELD_SEP + "ADJ")
|
|
|
|
|
|
def test_initialize_examples():
|
|
nlp = Language()
|
|
morphologizer = nlp.add_pipe("morphologizer")
|
|
morphologizer.add_label("POS" + Morphology.FIELD_SEP + "NOUN")
|
|
train_examples = []
|
|
for t in TRAIN_DATA:
|
|
train_examples.append(Example.from_dict(nlp.make_doc(t[0]), t[1]))
|
|
# you shouldn't really call this more than once, but for testing it should be fine
|
|
nlp.initialize()
|
|
nlp.initialize(get_examples=lambda: train_examples)
|
|
with pytest.raises(TypeError):
|
|
nlp.initialize(get_examples=lambda: None)
|
|
with pytest.raises(TypeError):
|
|
nlp.initialize(get_examples=train_examples)
|
|
|
|
|
|
def test_overfitting_IO():
|
|
# Simple test to try and quickly overfit the morphologizer - ensuring the ML models work correctly
|
|
nlp = English()
|
|
nlp.add_pipe("morphologizer")
|
|
train_examples = []
|
|
for inst in TRAIN_DATA:
|
|
train_examples.append(Example.from_dict(nlp.make_doc(inst[0]), inst[1]))
|
|
optimizer = nlp.initialize(get_examples=lambda: train_examples)
|
|
|
|
for i in range(50):
|
|
losses = {}
|
|
nlp.update(train_examples, sgd=optimizer, losses=losses)
|
|
assert losses["morphologizer"] < 0.00001
|
|
|
|
# test the trained model
|
|
test_text = "I like blue ham"
|
|
doc = nlp(test_text)
|
|
gold_morphs = ["Feat=N", "Feat=V", "", ""]
|
|
gold_pos_tags = ["NOUN", "VERB", "ADJ", ""]
|
|
assert [str(t.morph) for t in doc] == gold_morphs
|
|
assert [t.pos_ for t in doc] == gold_pos_tags
|
|
|
|
# Also test the results are still the same after IO
|
|
with make_tempdir() as tmp_dir:
|
|
nlp.to_disk(tmp_dir)
|
|
nlp2 = util.load_model_from_path(tmp_dir)
|
|
doc2 = nlp2(test_text)
|
|
assert [str(t.morph) for t in doc2] == gold_morphs
|
|
assert [t.pos_ for t in doc2] == gold_pos_tags
|
|
|
|
# Make sure that running pipe twice, or comparing to call, always amounts to the same predictions
|
|
texts = [
|
|
"Just a sentence.",
|
|
"Then one more sentence about London.",
|
|
"Here is another one.",
|
|
"I like London.",
|
|
]
|
|
batch_deps_1 = [doc.to_array([MORPH]) for doc in nlp.pipe(texts)]
|
|
batch_deps_2 = [doc.to_array([MORPH]) for doc in nlp.pipe(texts)]
|
|
no_batch_deps = [doc.to_array([MORPH]) for doc in [nlp(text) for text in texts]]
|
|
assert_equal(batch_deps_1, batch_deps_2)
|
|
assert_equal(batch_deps_1, no_batch_deps)
|
|
|
|
# Test without POS
|
|
nlp.remove_pipe("morphologizer")
|
|
nlp.add_pipe("morphologizer")
|
|
for example in train_examples:
|
|
for token in example.reference:
|
|
token.pos_ = ""
|
|
optimizer = nlp.initialize(get_examples=lambda: train_examples)
|
|
for i in range(50):
|
|
losses = {}
|
|
nlp.update(train_examples, sgd=optimizer, losses=losses)
|
|
assert losses["morphologizer"] < 0.00001
|
|
|
|
# Test the trained model
|
|
test_text = "I like blue ham"
|
|
doc = nlp(test_text)
|
|
gold_morphs = ["Feat=N", "Feat=V", "", ""]
|
|
gold_pos_tags = ["", "", "", ""]
|
|
assert [str(t.morph) for t in doc] == gold_morphs
|
|
assert [t.pos_ for t in doc] == gold_pos_tags
|
|
|
|
# Test overwrite+extend settings
|
|
# (note that "" is unset, "_" is set and empty)
|
|
morphs = ["Feat=V", "Feat=N", "_"]
|
|
doc = Doc(nlp.vocab, words=["blue", "ham", "like"], morphs=morphs)
|
|
orig_morphs = [str(t.morph) for t in doc]
|
|
orig_pos_tags = [t.pos_ for t in doc]
|
|
morphologizer = nlp.get_pipe("morphologizer")
|
|
|
|
# don't overwrite or extend
|
|
morphologizer.cfg["overwrite"] = False
|
|
doc = morphologizer(doc)
|
|
assert [str(t.morph) for t in doc] == orig_morphs
|
|
assert [t.pos_ for t in doc] == orig_pos_tags
|
|
|
|
# overwrite and extend
|
|
morphologizer.cfg["overwrite"] = True
|
|
morphologizer.cfg["extend"] = True
|
|
doc = Doc(nlp.vocab, words=["I", "like"], morphs=["Feat=A|That=A|This=A", ""])
|
|
doc = morphologizer(doc)
|
|
assert [str(t.morph) for t in doc] == ["Feat=N|That=A|This=A", "Feat=V"]
|
|
|
|
# extend without overwriting
|
|
morphologizer.cfg["overwrite"] = False
|
|
morphologizer.cfg["extend"] = True
|
|
doc = Doc(nlp.vocab, words=["I", "like"], morphs=["Feat=A|That=A|This=A", "That=B"])
|
|
doc = morphologizer(doc)
|
|
assert [str(t.morph) for t in doc] == ["Feat=A|That=A|This=A", "Feat=V|That=B"]
|
|
|
|
# overwrite without extending
|
|
morphologizer.cfg["overwrite"] = True
|
|
morphologizer.cfg["extend"] = False
|
|
doc = Doc(nlp.vocab, words=["I", "like"], morphs=["Feat=A|That=A|This=A", ""])
|
|
doc = morphologizer(doc)
|
|
assert [str(t.morph) for t in doc] == ["Feat=N", "Feat=V"]
|
|
|
|
# Test with unset morph and partial POS
|
|
nlp.remove_pipe("morphologizer")
|
|
nlp.add_pipe("morphologizer")
|
|
for example in train_examples:
|
|
for token in example.reference:
|
|
if token.text == "ham":
|
|
token.pos_ = "NOUN"
|
|
else:
|
|
token.pos_ = ""
|
|
token.set_morph(None)
|
|
optimizer = nlp.initialize(get_examples=lambda: train_examples)
|
|
print(nlp.get_pipe("morphologizer").labels)
|
|
for i in range(50):
|
|
losses = {}
|
|
nlp.update(train_examples, sgd=optimizer, losses=losses)
|
|
assert losses["morphologizer"] < 0.00001
|
|
|
|
# Test the trained model
|
|
test_text = "I like blue ham"
|
|
doc = nlp(test_text)
|
|
gold_morphs = ["", "", "", ""]
|
|
gold_pos_tags = ["NOUN", "NOUN", "NOUN", "NOUN"]
|
|
assert [str(t.morph) for t in doc] == gold_morphs
|
|
assert [t.pos_ for t in doc] == gold_pos_tags
|