2020-05-19 16:59:14 +03:00
|
|
|
import pickle
|
2021-12-04 22:34:48 +03:00
|
|
|
|
|
|
|
import pytest
|
2021-10-27 15:08:31 +03:00
|
|
|
from thinc.api import get_current_ops
|
2021-12-04 22:34:48 +03:00
|
|
|
|
|
|
|
import spacy
|
|
|
|
from spacy.lang.en import English
|
2018-07-25 00:38:44 +03:00
|
|
|
from spacy.strings import StringStore
|
2021-12-04 22:34:48 +03:00
|
|
|
from spacy.tokens import Doc
|
|
|
|
from spacy.util import ensure_path, load_model
|
2021-10-27 15:08:31 +03:00
|
|
|
from spacy.vectors import Vectors
|
2021-12-04 22:34:48 +03:00
|
|
|
from spacy.vocab import Vocab
|
2018-07-25 00:38:44 +03:00
|
|
|
|
|
|
|
from ..util import make_tempdir
|
2017-06-02 11:57:42 +03:00
|
|
|
|
2023-03-20 01:41:20 +03:00
|
|
|
test_strings = [
|
|
|
|
(StringStore(), StringStore()),
|
|
|
|
(StringStore(["rats", "are", "cute"]), StringStore(["i", "like", "rats"])),
|
|
|
|
]
|
|
|
|
test_strings_attrs = [(StringStore(["rats", "are", "cute"]), "Hello")]
|
2017-06-02 11:57:42 +03:00
|
|
|
|
|
|
|
|
2021-12-04 22:34:48 +03:00
|
|
|
@pytest.mark.issue(599)
|
|
|
|
def test_issue599(en_vocab):
|
|
|
|
doc = Doc(en_vocab)
|
|
|
|
doc2 = Doc(doc.vocab)
|
|
|
|
doc2.from_bytes(doc.to_bytes())
|
|
|
|
assert doc2.has_annotation("DEP")
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.issue(4054)
|
|
|
|
def test_issue4054(en_vocab):
|
|
|
|
"""Test that a new blank model can be made with a vocab from file,
|
|
|
|
and that serialization does not drop the language at any point."""
|
|
|
|
nlp1 = English()
|
|
|
|
vocab1 = nlp1.vocab
|
|
|
|
with make_tempdir() as d:
|
|
|
|
vocab_dir = ensure_path(d / "vocab")
|
|
|
|
if not vocab_dir.exists():
|
|
|
|
vocab_dir.mkdir()
|
|
|
|
vocab1.to_disk(vocab_dir)
|
|
|
|
vocab2 = Vocab().from_disk(vocab_dir)
|
|
|
|
nlp2 = spacy.blank("en", vocab=vocab2)
|
|
|
|
nlp_dir = ensure_path(d / "nlp")
|
|
|
|
if not nlp_dir.exists():
|
|
|
|
nlp_dir.mkdir()
|
|
|
|
nlp2.to_disk(nlp_dir)
|
|
|
|
nlp3 = load_model(nlp_dir)
|
|
|
|
assert nlp3.lang == "en"
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.mark.issue(4133)
|
|
|
|
def test_issue4133(en_vocab):
|
|
|
|
nlp = English()
|
|
|
|
vocab_bytes = nlp.vocab.to_bytes()
|
|
|
|
words = ["Apple", "is", "looking", "at", "buying", "a", "startup"]
|
|
|
|
pos = ["NOUN", "VERB", "ADP", "VERB", "PROPN", "NOUN", "ADP"]
|
|
|
|
doc = Doc(en_vocab, words=words)
|
|
|
|
for i, token in enumerate(doc):
|
|
|
|
token.pos_ = pos[i]
|
|
|
|
# usually this is already True when starting from proper models instead of blank English
|
|
|
|
doc_bytes = doc.to_bytes()
|
|
|
|
vocab = Vocab()
|
|
|
|
vocab = vocab.from_bytes(vocab_bytes)
|
|
|
|
doc = Doc(vocab).from_bytes(doc_bytes)
|
|
|
|
actual = []
|
|
|
|
for token in doc:
|
|
|
|
actual.append(token.pos_)
|
|
|
|
assert actual == pos
|
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("text", ["rat"])
|
2018-07-25 00:38:44 +03:00
|
|
|
def test_serialize_vocab(en_vocab, text):
|
|
|
|
text_hash = en_vocab.strings.add(text)
|
2019-09-09 20:17:55 +03:00
|
|
|
vocab_bytes = en_vocab.to_bytes(exclude=["lookups"])
|
2018-07-25 00:38:44 +03:00
|
|
|
new_vocab = Vocab().from_bytes(vocab_bytes)
|
2019-03-10 18:36:29 +03:00
|
|
|
assert new_vocab.strings[text_hash] == text
|
2019-09-09 20:17:55 +03:00
|
|
|
assert new_vocab.to_bytes(exclude=["lookups"]) == vocab_bytes
|
2018-07-25 00:38:44 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings1,strings2", test_strings)
|
2018-07-25 00:38:44 +03:00
|
|
|
def test_serialize_vocab_roundtrip_bytes(strings1, strings2):
|
2017-06-02 11:57:42 +03:00
|
|
|
vocab1 = Vocab(strings=strings1)
|
|
|
|
vocab2 = Vocab(strings=strings2)
|
|
|
|
vocab1_b = vocab1.to_bytes()
|
|
|
|
vocab2_b = vocab2.to_bytes()
|
2023-03-20 01:41:20 +03:00
|
|
|
if strings1.to_bytes() == strings2.to_bytes():
|
2017-06-02 11:57:42 +03:00
|
|
|
assert vocab1_b == vocab2_b
|
|
|
|
else:
|
|
|
|
assert vocab1_b != vocab2_b
|
|
|
|
vocab1 = vocab1.from_bytes(vocab1_b)
|
|
|
|
assert vocab1.to_bytes() == vocab1_b
|
|
|
|
new_vocab1 = Vocab().from_bytes(vocab1_b)
|
|
|
|
assert new_vocab1.to_bytes() == vocab1_b
|
2020-08-07 16:27:13 +03:00
|
|
|
assert len(new_vocab1.strings) == len(strings1)
|
|
|
|
assert sorted([s for s in new_vocab1.strings]) == sorted(strings1)
|
2017-06-02 11:57:42 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings1,strings2", test_strings)
|
|
|
|
def test_serialize_vocab_roundtrip_disk(strings1, strings2):
|
2017-06-02 11:57:42 +03:00
|
|
|
vocab1 = Vocab(strings=strings1)
|
|
|
|
vocab2 = Vocab(strings=strings2)
|
|
|
|
with make_tempdir() as d:
|
2018-11-27 03:09:36 +03:00
|
|
|
file_path1 = d / "vocab1"
|
|
|
|
file_path2 = d / "vocab2"
|
2017-06-02 11:57:42 +03:00
|
|
|
vocab1.to_disk(file_path1)
|
|
|
|
vocab2.to_disk(file_path2)
|
|
|
|
vocab1_d = Vocab().from_disk(file_path1)
|
|
|
|
vocab2_d = Vocab().from_disk(file_path2)
|
2020-05-19 16:59:14 +03:00
|
|
|
# check strings rather than lexemes, which are only reloaded on demand
|
2021-04-09 12:53:13 +03:00
|
|
|
assert set(strings1) == set([s for s in vocab1_d.strings])
|
|
|
|
assert set(strings2) == set([s for s in vocab2_d.strings])
|
|
|
|
if set(strings1) == set(strings2):
|
2020-08-07 16:27:13 +03:00
|
|
|
assert [s for s in vocab1_d.strings] == [s for s in vocab2_d.strings]
|
2017-06-02 11:57:42 +03:00
|
|
|
else:
|
2020-08-07 16:27:13 +03:00
|
|
|
assert [s for s in vocab1_d.strings] != [s for s in vocab2_d.strings]
|
2017-06-02 11:57:42 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings,lex_attr", test_strings_attrs)
|
2017-06-02 11:57:42 +03:00
|
|
|
def test_serialize_vocab_lex_attrs_bytes(strings, lex_attr):
|
|
|
|
vocab1 = Vocab(strings=strings)
|
|
|
|
vocab2 = Vocab()
|
2023-03-20 01:41:20 +03:00
|
|
|
s = next(iter(vocab1.strings))
|
|
|
|
vocab1[s].norm_ = lex_attr
|
|
|
|
assert vocab1[s].norm_ == lex_attr
|
|
|
|
assert vocab2[s].norm_ != lex_attr
|
2017-06-02 11:57:42 +03:00
|
|
|
vocab2 = vocab2.from_bytes(vocab1.to_bytes())
|
2023-03-20 01:41:20 +03:00
|
|
|
assert vocab2[s].norm_ == lex_attr
|
2017-06-02 11:57:42 +03:00
|
|
|
|
|
|
|
|
2019-03-10 18:36:29 +03:00
|
|
|
@pytest.mark.parametrize("strings,lex_attr", test_strings_attrs)
|
|
|
|
def test_deserialize_vocab_seen_entries(strings, lex_attr):
|
|
|
|
# Reported in #2153
|
|
|
|
vocab = Vocab(strings=strings)
|
|
|
|
vocab.from_bytes(vocab.to_bytes())
|
2020-08-07 16:27:13 +03:00
|
|
|
assert len(vocab.strings) == len(strings)
|
2019-03-10 18:36:29 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings,lex_attr", test_strings_attrs)
|
2017-06-02 11:57:42 +03:00
|
|
|
def test_serialize_vocab_lex_attrs_disk(strings, lex_attr):
|
|
|
|
vocab1 = Vocab(strings=strings)
|
|
|
|
vocab2 = Vocab()
|
2023-03-20 01:41:20 +03:00
|
|
|
s = next(iter(vocab1.strings))
|
|
|
|
vocab1[s].norm_ = lex_attr
|
|
|
|
assert vocab1[s].norm_ == lex_attr
|
|
|
|
assert vocab2[s].norm_ != lex_attr
|
2017-06-02 11:57:42 +03:00
|
|
|
with make_tempdir() as d:
|
2018-11-27 03:09:36 +03:00
|
|
|
file_path = d / "vocab"
|
2017-06-02 11:57:42 +03:00
|
|
|
vocab1.to_disk(file_path)
|
|
|
|
vocab2 = vocab2.from_disk(file_path)
|
2023-03-20 01:41:20 +03:00
|
|
|
assert vocab2[s].norm_ == lex_attr
|
2018-07-25 00:38:44 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings1,strings2", test_strings)
|
2018-07-25 00:38:44 +03:00
|
|
|
def test_serialize_stringstore_roundtrip_bytes(strings1, strings2):
|
|
|
|
sstore1 = StringStore(strings=strings1)
|
|
|
|
sstore2 = StringStore(strings=strings2)
|
|
|
|
sstore1_b = sstore1.to_bytes()
|
|
|
|
sstore2_b = sstore2.to_bytes()
|
2021-04-09 12:53:13 +03:00
|
|
|
if set(strings1) == set(strings2):
|
2018-07-25 00:38:44 +03:00
|
|
|
assert sstore1_b == sstore2_b
|
|
|
|
else:
|
|
|
|
assert sstore1_b != sstore2_b
|
|
|
|
sstore1 = sstore1.from_bytes(sstore1_b)
|
|
|
|
assert sstore1.to_bytes() == sstore1_b
|
|
|
|
new_sstore1 = StringStore().from_bytes(sstore1_b)
|
|
|
|
assert new_sstore1.to_bytes() == sstore1_b
|
2021-04-09 12:53:13 +03:00
|
|
|
assert set(new_sstore1) == set(strings1)
|
2018-07-25 00:38:44 +03:00
|
|
|
|
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
@pytest.mark.parametrize("strings1,strings2", test_strings)
|
2018-07-25 00:38:44 +03:00
|
|
|
def test_serialize_stringstore_roundtrip_disk(strings1, strings2):
|
|
|
|
sstore1 = StringStore(strings=strings1)
|
|
|
|
sstore2 = StringStore(strings=strings2)
|
|
|
|
with make_tempdir() as d:
|
2018-11-27 03:09:36 +03:00
|
|
|
file_path1 = d / "strings1"
|
|
|
|
file_path2 = d / "strings2"
|
2018-07-25 00:38:44 +03:00
|
|
|
sstore1.to_disk(file_path1)
|
|
|
|
sstore2.to_disk(file_path2)
|
|
|
|
sstore1_d = StringStore().from_disk(file_path1)
|
|
|
|
sstore2_d = StringStore().from_disk(file_path2)
|
2021-04-09 12:53:13 +03:00
|
|
|
assert set(sstore1_d) == set(sstore1)
|
|
|
|
assert set(sstore2_d) == set(sstore2)
|
|
|
|
if set(strings1) == set(strings2):
|
|
|
|
assert set(sstore1_d) == set(sstore2_d)
|
2018-07-25 00:38:44 +03:00
|
|
|
else:
|
2021-04-09 12:53:13 +03:00
|
|
|
assert set(sstore1_d) != set(sstore2_d)
|
2020-05-19 16:59:14 +03:00
|
|
|
|
2020-05-21 15:14:01 +03:00
|
|
|
|
2020-05-19 16:59:14 +03:00
|
|
|
@pytest.mark.parametrize("strings,lex_attr", test_strings_attrs)
|
|
|
|
def test_pickle_vocab(strings, lex_attr):
|
|
|
|
vocab = Vocab(strings=strings)
|
2021-10-27 15:08:31 +03:00
|
|
|
ops = get_current_ops()
|
|
|
|
vectors = Vectors(data=ops.xp.zeros((10, 10)), mode="floret", hash_count=1)
|
|
|
|
vocab.vectors = vectors
|
2020-05-19 16:59:14 +03:00
|
|
|
vocab[strings[0]].norm_ = lex_attr
|
|
|
|
vocab_pickled = pickle.dumps(vocab)
|
|
|
|
vocab_unpickled = pickle.loads(vocab_pickled)
|
|
|
|
assert vocab.to_bytes() == vocab_unpickled.to_bytes()
|
2021-10-27 15:08:31 +03:00
|
|
|
assert vocab_unpickled.vectors.mode == "floret"
|