2017-01-12 17:30:49 +03:00
|
|
|
# coding: utf-8
|
2015-04-07 05:52:25 +03:00
|
|
|
from __future__ import unicode_literals
|
2017-01-12 17:30:49 +03:00
|
|
|
|
2015-04-07 05:52:25 +03:00
|
|
|
import pytest
|
2021-05-17 17:54:10 +03:00
|
|
|
import numpy
|
|
|
|
from numpy.testing import assert_array_equal
|
2018-07-25 00:38:44 +03:00
|
|
|
from spacy.attrs import ORTH, LENGTH
|
2018-12-08 15:08:41 +03:00
|
|
|
from spacy.tokens import Doc, Span
|
2018-07-25 00:38:44 +03:00
|
|
|
from spacy.vocab import Vocab
|
2019-05-08 03:33:40 +03:00
|
|
|
from spacy.util import filter_spans
|
2018-07-25 00:38:44 +03:00
|
|
|
|
|
|
|
from ..util import get_doc
|
2015-04-07 05:52:25 +03:00
|
|
|
|
|
|
|
|
|
|
|
@pytest.fixture
|
2017-01-12 17:30:49 +03:00
|
|
|
def doc(en_tokenizer):
|
2018-11-27 03:09:36 +03:00
|
|
|
# fmt: off
|
2017-01-12 17:30:49 +03:00
|
|
|
text = "This is a sentence. This is another sentence. And a third."
|
|
|
|
heads = [1, 0, 1, -2, -3, 1, 0, 1, -2, -3, 0, 1, -2, -1]
|
2018-11-27 03:09:36 +03:00
|
|
|
deps = ["nsubj", "ROOT", "det", "attr", "punct", "nsubj", "ROOT", "det",
|
|
|
|
"attr", "punct", "ROOT", "det", "npadvmod", "punct"]
|
|
|
|
# fmt: on
|
2017-01-12 17:30:49 +03:00
|
|
|
tokens = en_tokenizer(text)
|
2018-07-25 00:38:44 +03:00
|
|
|
return get_doc(tokens.vocab, words=[t.text for t in tokens], heads=heads, deps=deps)
|
2015-04-07 05:52:25 +03:00
|
|
|
|
|
|
|
|
2018-03-27 20:23:02 +03:00
|
|
|
@pytest.fixture
|
|
|
|
def doc_not_parsed(en_tokenizer):
|
|
|
|
text = "This is a sentence. This is another sentence. And a third."
|
|
|
|
tokens = en_tokenizer(text)
|
2018-07-25 00:38:44 +03:00
|
|
|
doc = Doc(tokens.vocab, words=[t.text for t in tokens])
|
|
|
|
doc.is_parsed = False
|
|
|
|
return doc
|
2018-03-27 20:23:02 +03:00
|
|
|
|
|
|
|
|
2019-12-13 17:54:58 +03:00
|
|
|
@pytest.mark.parametrize(
|
|
|
|
"i_sent,i,j,text",
|
|
|
|
[
|
|
|
|
(0, 0, len("This is a"), "This is a"),
|
|
|
|
(1, 0, len("This is another"), "This is another"),
|
|
|
|
(2, len("And "), len("And ") + len("a third"), "a third"),
|
|
|
|
(0, 1, 2, None),
|
|
|
|
],
|
|
|
|
)
|
|
|
|
def test_char_span(doc, i_sent, i, j, text):
|
|
|
|
sents = list(doc.sents)
|
|
|
|
span = sents[i_sent].char_span(i, j)
|
|
|
|
if not text:
|
|
|
|
assert not span
|
|
|
|
else:
|
|
|
|
assert span.text == text
|
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_sent_spans(doc):
|
2015-04-07 05:52:25 +03:00
|
|
|
sents = list(doc.sents)
|
|
|
|
assert sents[0].start == 0
|
|
|
|
assert sents[0].end == 5
|
|
|
|
assert len(sents) == 3
|
|
|
|
assert sum(len(sent) for sent in sents) == len(doc)
|
2015-07-09 18:30:58 +03:00
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_root(doc):
|
|
|
|
span = doc[2:4]
|
|
|
|
assert len(span) == 2
|
2018-11-27 03:09:36 +03:00
|
|
|
assert span.text == "a sentence"
|
|
|
|
assert span.root.text == "sentence"
|
|
|
|
assert span.root.head.text == "is"
|
2016-01-16 18:19:09 +03:00
|
|
|
|
2018-03-27 20:23:02 +03:00
|
|
|
|
2017-03-11 03:50:02 +03:00
|
|
|
def test_spans_string_fn(doc):
|
|
|
|
span = doc[0:4]
|
|
|
|
assert len(span) == 4
|
2018-11-27 03:09:36 +03:00
|
|
|
assert span.text == "This is a sentence"
|
|
|
|
assert span.upper_ == "THIS IS A SENTENCE"
|
|
|
|
assert span.lower_ == "this is a sentence"
|
2016-01-16 18:19:09 +03:00
|
|
|
|
2018-03-27 20:23:02 +03:00
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_root2(en_tokenizer):
|
|
|
|
text = "through North and South Carolina"
|
|
|
|
heads = [0, 3, -1, -2, -4]
|
|
|
|
tokens = en_tokenizer(text)
|
2018-07-25 00:38:44 +03:00
|
|
|
doc = get_doc(tokens.vocab, words=[t.text for t in tokens], heads=heads)
|
2018-11-27 03:09:36 +03:00
|
|
|
assert doc[-2:].root.text == "Carolina"
|
2016-05-06 01:17:38 +03:00
|
|
|
|
|
|
|
|
2018-03-27 20:23:02 +03:00
|
|
|
def test_spans_span_sent(doc, doc_not_parsed):
|
2017-01-12 17:30:49 +03:00
|
|
|
"""Test span.sent property"""
|
2016-05-06 01:17:38 +03:00
|
|
|
assert len(list(doc.sents))
|
2018-11-27 03:09:36 +03:00
|
|
|
assert doc[:2].sent.root.text == "is"
|
|
|
|
assert doc[:2].sent.text == "This is a sentence ."
|
|
|
|
assert doc[6:7].sent.root.left_edge.text == "This"
|
2018-03-27 20:23:02 +03:00
|
|
|
# test on manual sbd
|
|
|
|
doc_not_parsed[0].is_sent_start = True
|
|
|
|
doc_not_parsed[5].is_sent_start = True
|
|
|
|
assert doc_not_parsed[1:3].sent == doc_not_parsed[0:5]
|
|
|
|
assert doc_not_parsed[10:14].sent == doc_not_parsed[5:]
|
2017-01-12 17:30:49 +03:00
|
|
|
|
|
|
|
|
2017-10-20 21:28:00 +03:00
|
|
|
def test_spans_lca_matrix(en_tokenizer):
|
|
|
|
"""Test span's lca matrix generation"""
|
2018-11-27 03:09:36 +03:00
|
|
|
tokens = en_tokenizer("the lazy dog slept")
|
2018-07-25 00:38:44 +03:00
|
|
|
doc = get_doc(tokens.vocab, words=[t.text for t in tokens], heads=[2, 1, 1, 0])
|
2017-10-20 21:28:00 +03:00
|
|
|
lca = doc[:2].get_lca_matrix()
|
2019-01-06 21:07:50 +03:00
|
|
|
assert lca.shape == (2, 2)
|
|
|
|
assert lca[0, 0] == 0 # the & the -> the
|
2019-02-07 22:54:07 +03:00
|
|
|
assert lca[0, 1] == -1 # the & lazy -> dog (out of span)
|
|
|
|
assert lca[1, 0] == -1 # lazy & the -> dog (out of span)
|
2019-01-06 21:07:50 +03:00
|
|
|
assert lca[1, 1] == 1 # lazy & lazy -> lazy
|
|
|
|
|
|
|
|
lca = doc[1:].get_lca_matrix()
|
|
|
|
assert lca.shape == (3, 3)
|
2019-02-07 22:54:07 +03:00
|
|
|
assert lca[0, 0] == 0 # lazy & lazy -> lazy
|
|
|
|
assert lca[0, 1] == 1 # lazy & dog -> dog
|
|
|
|
assert lca[0, 2] == 2 # lazy & slept -> slept
|
2019-01-06 21:07:50 +03:00
|
|
|
|
|
|
|
lca = doc[2:].get_lca_matrix()
|
|
|
|
assert lca.shape == (2, 2)
|
2019-02-07 22:54:07 +03:00
|
|
|
assert lca[0, 0] == 0 # dog & dog -> dog
|
|
|
|
assert lca[0, 1] == 1 # dog & slept -> slept
|
|
|
|
assert lca[1, 0] == 1 # slept & dog -> slept
|
|
|
|
assert lca[1, 1] == 1 # slept & slept -> slept
|
2017-10-20 21:28:00 +03:00
|
|
|
|
2021-05-17 17:54:10 +03:00
|
|
|
# example from Span API docs
|
|
|
|
tokens = en_tokenizer("I like New York in Autumn")
|
|
|
|
doc = get_doc(
|
|
|
|
tokens.vocab, words=[t.text for t in tokens], heads=[1, 0, 1, -2, -1, -1]
|
|
|
|
)
|
|
|
|
lca = doc[1:4].get_lca_matrix()
|
|
|
|
assert_array_equal(lca, numpy.asarray([[0, 0, 0], [0, 1, 2], [0, 2, 2]]))
|
|
|
|
|
2017-10-20 21:28:00 +03:00
|
|
|
|
2018-01-15 18:29:48 +03:00
|
|
|
def test_span_similarity_match():
|
2018-11-27 03:09:36 +03:00
|
|
|
doc = Doc(Vocab(), words=["a", "b", "a", "b"])
|
2018-01-15 18:29:48 +03:00
|
|
|
span1 = doc[:2]
|
|
|
|
span2 = doc[2:]
|
2020-04-28 14:37:37 +03:00
|
|
|
with pytest.warns(UserWarning):
|
2018-05-21 02:22:38 +03:00
|
|
|
assert span1.similarity(span2) == 1.0
|
|
|
|
assert span1.similarity(doc) == 0.0
|
2018-11-27 03:09:36 +03:00
|
|
|
assert span1[:1].similarity(doc.vocab["a"]) == 1.0
|
2018-01-15 18:29:48 +03:00
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_default_sentiment(en_tokenizer):
|
|
|
|
"""Test span.sentiment property's default averaging behaviour"""
|
|
|
|
text = "good stuff bad stuff"
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
tokens.vocab[tokens[0].text].sentiment = 3.0
|
|
|
|
tokens.vocab[tokens[2].text].sentiment = -2.0
|
2018-07-25 00:38:44 +03:00
|
|
|
doc = Doc(tokens.vocab, words=[t.text for t in tokens])
|
2017-01-12 17:30:49 +03:00
|
|
|
assert doc[:2].sentiment == 3.0 / 2
|
2018-11-27 03:09:36 +03:00
|
|
|
assert doc[-2:].sentiment == -2.0 / 2
|
|
|
|
assert doc[:-1].sentiment == (3.0 + -2) / 3.0
|
2017-01-12 17:30:49 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_spans_override_sentiment(en_tokenizer):
|
|
|
|
"""Test span.sentiment property's default averaging behaviour"""
|
|
|
|
text = "good stuff bad stuff"
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
tokens.vocab[tokens[0].text].sentiment = 3.0
|
|
|
|
tokens.vocab[tokens[2].text].sentiment = -2.0
|
2018-07-25 00:38:44 +03:00
|
|
|
doc = Doc(tokens.vocab, words=[t.text for t in tokens])
|
2018-11-27 03:09:36 +03:00
|
|
|
doc.user_span_hooks["sentiment"] = lambda span: 10.0
|
2017-01-12 17:30:49 +03:00
|
|
|
assert doc[:2].sentiment == 10.0
|
|
|
|
assert doc[-2:].sentiment == 10.0
|
|
|
|
assert doc[:-1].sentiment == 10.0
|
2017-04-26 20:01:05 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_spans_are_hashable(en_tokenizer):
|
|
|
|
"""Test spans can be hashed."""
|
|
|
|
text = "good stuff bad stuff"
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
span1 = tokens[:2]
|
|
|
|
span2 = tokens[2:4]
|
|
|
|
assert hash(span1) != hash(span2)
|
|
|
|
span3 = tokens[0:2]
|
|
|
|
assert hash(span3) == hash(span1)
|
2017-10-24 16:27:29 +03:00
|
|
|
|
2017-08-19 17:18:23 +03:00
|
|
|
|
|
|
|
def test_spans_by_character(doc):
|
|
|
|
span1 = doc[1:-2]
|
2020-08-04 14:36:32 +03:00
|
|
|
|
|
|
|
# default and specified alignment mode "strict"
|
2018-11-27 03:09:36 +03:00
|
|
|
span2 = doc.char_span(span1.start_char, span1.end_char, label="GPE")
|
2017-08-19 17:18:23 +03:00
|
|
|
assert span1.start_char == span2.start_char
|
|
|
|
assert span1.end_char == span2.end_char
|
2018-11-27 03:09:36 +03:00
|
|
|
assert span2.label_ == "GPE"
|
2017-08-19 17:24:38 +03:00
|
|
|
|
2020-10-01 15:01:52 +03:00
|
|
|
span2 = doc.char_span(
|
|
|
|
span1.start_char, span1.end_char, label="GPE", alignment_mode="strict"
|
|
|
|
)
|
2020-08-04 14:36:32 +03:00
|
|
|
assert span1.start_char == span2.start_char
|
|
|
|
assert span1.end_char == span2.end_char
|
|
|
|
assert span2.label_ == "GPE"
|
|
|
|
|
|
|
|
# alignment mode "contract"
|
2020-10-01 15:01:52 +03:00
|
|
|
span2 = doc.char_span(
|
|
|
|
span1.start_char - 3, span1.end_char, label="GPE", alignment_mode="contract"
|
|
|
|
)
|
2020-08-04 14:36:32 +03:00
|
|
|
assert span1.start_char == span2.start_char
|
|
|
|
assert span1.end_char == span2.end_char
|
|
|
|
assert span2.label_ == "GPE"
|
|
|
|
|
|
|
|
# alignment mode "expand"
|
2020-10-01 15:01:52 +03:00
|
|
|
span2 = doc.char_span(
|
|
|
|
span1.start_char + 1, span1.end_char, label="GPE", alignment_mode="expand"
|
|
|
|
)
|
2020-08-04 14:36:32 +03:00
|
|
|
assert span1.start_char == span2.start_char
|
|
|
|
assert span1.end_char == span2.end_char
|
|
|
|
assert span2.label_ == "GPE"
|
|
|
|
|
2021-01-27 15:40:42 +03:00
|
|
|
# unsupported alignment mode
|
|
|
|
with pytest.raises(ValueError):
|
|
|
|
span2 = doc.char_span(
|
|
|
|
span1.start_char + 1, span1.end_char, label="GPE", alignment_mode="unk"
|
|
|
|
)
|
|
|
|
|
2017-08-19 17:24:38 +03:00
|
|
|
|
|
|
|
def test_span_to_array(doc):
|
|
|
|
span = doc[1:-2]
|
|
|
|
arr = span.to_array([ORTH, LENGTH])
|
|
|
|
assert arr.shape == (len(span), 2)
|
|
|
|
assert arr[0, 0] == span[0].orth
|
|
|
|
assert arr[0, 1] == len(span[0])
|
|
|
|
|
2017-10-24 16:28:05 +03:00
|
|
|
|
2018-07-25 00:38:44 +03:00
|
|
|
def test_span_as_doc(doc):
|
|
|
|
span = doc[4:10]
|
|
|
|
span_doc = span.as_doc()
|
|
|
|
assert span.text == span_doc.text.strip()
|
2018-12-30 17:17:46 +03:00
|
|
|
assert isinstance(span_doc, doc.__class__)
|
|
|
|
assert span_doc is not doc
|
|
|
|
assert span_doc[0].idx == 0
|
2018-08-07 14:52:32 +03:00
|
|
|
|
2018-11-27 03:09:36 +03:00
|
|
|
|
2019-09-12 18:08:14 +03:00
|
|
|
def test_span_as_doc_user_data(doc):
|
|
|
|
"""Test that the user_data can be preserved (but not by default). """
|
|
|
|
my_key = "my_info"
|
|
|
|
my_value = 342
|
|
|
|
doc.user_data[my_key] = my_value
|
|
|
|
|
|
|
|
span = doc[4:10]
|
|
|
|
span_doc_with = span.as_doc(copy_user_data=True)
|
|
|
|
span_doc_without = span.as_doc()
|
|
|
|
|
|
|
|
assert doc.user_data.get(my_key, None) is my_value
|
|
|
|
assert span_doc_with.user_data.get(my_key, None) is my_value
|
|
|
|
assert span_doc_without.user_data.get(my_key, None) is None
|
|
|
|
|
|
|
|
|
2019-03-22 14:05:35 +03:00
|
|
|
def test_span_string_label_kb_id(doc):
|
|
|
|
span = Span(doc, 0, 1, label="hello", kb_id="Q342")
|
2019-02-07 22:54:07 +03:00
|
|
|
assert span.label_ == "hello"
|
|
|
|
assert span.label == doc.vocab.strings["hello"]
|
2019-03-22 14:05:35 +03:00
|
|
|
assert span.kb_id_ == "Q342"
|
|
|
|
assert span.kb_id == doc.vocab.strings["Q342"]
|
2019-02-07 22:54:07 +03:00
|
|
|
|
2018-12-08 15:08:41 +03:00
|
|
|
|
2019-03-15 02:46:45 +03:00
|
|
|
def test_span_label_readonly(doc):
|
2018-12-08 15:08:41 +03:00
|
|
|
span = Span(doc, 0, 1)
|
2019-03-15 02:46:45 +03:00
|
|
|
with pytest.raises(NotImplementedError):
|
|
|
|
span.label_ = "hello"
|
2019-02-07 22:54:07 +03:00
|
|
|
|
2018-12-08 15:08:41 +03:00
|
|
|
|
2019-03-22 14:05:35 +03:00
|
|
|
def test_span_kb_id_readonly(doc):
|
|
|
|
span = Span(doc, 0, 1)
|
|
|
|
with pytest.raises(NotImplementedError):
|
|
|
|
span.kb_id_ = "Q342"
|
|
|
|
|
|
|
|
|
2018-08-07 14:52:32 +03:00
|
|
|
def test_span_ents_property(doc):
|
|
|
|
"""Test span.ents for the """
|
|
|
|
doc.ents = [
|
2018-11-27 03:09:36 +03:00
|
|
|
(doc.vocab.strings["PRODUCT"], 0, 1),
|
|
|
|
(doc.vocab.strings["PRODUCT"], 7, 8),
|
|
|
|
(doc.vocab.strings["PRODUCT"], 11, 14),
|
2018-08-07 14:52:32 +03:00
|
|
|
]
|
|
|
|
assert len(list(doc.ents)) == 3
|
|
|
|
sentences = list(doc.sents)
|
|
|
|
assert len(sentences) == 3
|
|
|
|
assert len(sentences[0].ents) == 1
|
|
|
|
# First sentence, also tests start of sentence
|
|
|
|
assert sentences[0].ents[0].text == "This"
|
|
|
|
assert sentences[0].ents[0].label_ == "PRODUCT"
|
|
|
|
assert sentences[0].ents[0].start == 0
|
|
|
|
assert sentences[0].ents[0].end == 1
|
|
|
|
# Second sentence
|
|
|
|
assert len(sentences[1].ents) == 1
|
|
|
|
assert sentences[1].ents[0].text == "another"
|
|
|
|
assert sentences[1].ents[0].label_ == "PRODUCT"
|
|
|
|
assert sentences[1].ents[0].start == 7
|
|
|
|
assert sentences[1].ents[0].end == 8
|
|
|
|
# Third sentence ents, Also tests end of sentence
|
|
|
|
assert sentences[2].ents[0].text == "a third ."
|
|
|
|
assert sentences[2].ents[0].label_ == "PRODUCT"
|
|
|
|
assert sentences[2].ents[0].start == 11
|
|
|
|
assert sentences[2].ents[0].end == 14
|
2019-05-08 03:33:40 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_filter_spans(doc):
|
|
|
|
# Test filtering duplicates
|
|
|
|
spans = [doc[1:4], doc[6:8], doc[1:4], doc[10:14]]
|
|
|
|
filtered = filter_spans(spans)
|
|
|
|
assert len(filtered) == 3
|
|
|
|
assert filtered[0].start == 1 and filtered[0].end == 4
|
|
|
|
assert filtered[1].start == 6 and filtered[1].end == 8
|
|
|
|
assert filtered[2].start == 10 and filtered[2].end == 14
|
|
|
|
# Test filtering overlaps with longest preference
|
|
|
|
spans = [doc[1:4], doc[1:3], doc[5:10], doc[7:9], doc[1:4]]
|
|
|
|
filtered = filter_spans(spans)
|
|
|
|
assert len(filtered) == 2
|
|
|
|
assert len(filtered[0]) == 3
|
|
|
|
assert len(filtered[1]) == 5
|
|
|
|
assert filtered[0].start == 1 and filtered[0].end == 4
|
|
|
|
assert filtered[1].start == 5 and filtered[1].end == 10
|
2019-10-10 18:00:03 +03:00
|
|
|
# Test filtering overlaps with earlier preference for identical length
|
|
|
|
spans = [doc[1:4], doc[2:5], doc[5:10], doc[7:9], doc[1:4]]
|
|
|
|
filtered = filter_spans(spans)
|
|
|
|
assert len(filtered) == 2
|
|
|
|
assert len(filtered[0]) == 3
|
|
|
|
assert len(filtered[1]) == 5
|
|
|
|
assert filtered[0].start == 1 and filtered[0].end == 4
|
|
|
|
assert filtered[1].start == 5 and filtered[1].end == 10
|
2020-02-16 19:20:36 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_span_eq_hash(doc, doc_not_parsed):
|
|
|
|
assert doc[0:2] == doc[0:2]
|
|
|
|
assert doc[0:2] != doc[1:3]
|
|
|
|
assert doc[0:2] != doc_not_parsed[0:2]
|
|
|
|
assert hash(doc[0:2]) == hash(doc[0:2])
|
|
|
|
assert hash(doc[0:2]) != hash(doc[1:3])
|
|
|
|
assert hash(doc[0:2]) != hash(doc_not_parsed[0:2])
|
2020-08-04 14:35:25 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_span_boundaries(doc):
|
|
|
|
start = 1
|
|
|
|
end = 5
|
|
|
|
span = doc[start:end]
|
|
|
|
for i in range(start, end):
|
|
|
|
assert span[i - start] == doc[i]
|
|
|
|
with pytest.raises(IndexError):
|
|
|
|
_ = span[-5]
|
|
|
|
with pytest.raises(IndexError):
|
|
|
|
_ = span[5]
|
2020-10-01 15:01:52 +03:00
|
|
|
|
|
|
|
|
|
|
|
def test_sent(en_tokenizer):
|
|
|
|
doc = en_tokenizer("Check span.sent raises error if doc is not sentencized.")
|
|
|
|
span = doc[1:3]
|
|
|
|
assert not span.doc.is_sentenced
|
|
|
|
with pytest.raises(ValueError):
|
|
|
|
span.sent
|