2017-01-12 17:30:49 +03:00
|
|
|
# coding: utf-8
|
2015-04-07 05:52:25 +03:00
|
|
|
from __future__ import unicode_literals
|
2017-01-12 17:30:49 +03:00
|
|
|
|
|
|
|
from ..util import get_doc
|
2015-04-07 05:52:25 +03:00
|
|
|
|
|
|
|
import pytest
|
|
|
|
|
|
|
|
|
|
|
|
@pytest.fixture
|
2017-01-12 17:30:49 +03:00
|
|
|
def doc(en_tokenizer):
|
|
|
|
text = "This is a sentence. This is another sentence. And a third."
|
|
|
|
heads = [1, 0, 1, -2, -3, 1, 0, 1, -2, -3, 0, 1, -2, -1]
|
|
|
|
deps = ['nsubj', 'ROOT', 'det', 'attr', 'punct', 'nsubj', 'ROOT', 'det',
|
|
|
|
'attr', 'punct', 'ROOT', 'det', 'npadvmod', 'punct']
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
return get_doc(tokens.vocab, [t.text for t in tokens], heads=heads, deps=deps)
|
2015-04-07 05:52:25 +03:00
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_sent_spans(doc):
|
2015-04-07 05:52:25 +03:00
|
|
|
sents = list(doc.sents)
|
|
|
|
assert sents[0].start == 0
|
|
|
|
assert sents[0].end == 5
|
|
|
|
assert len(sents) == 3
|
|
|
|
assert sum(len(sent) for sent in sents) == len(doc)
|
2015-07-09 18:30:58 +03:00
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_root(doc):
|
|
|
|
span = doc[2:4]
|
|
|
|
assert len(span) == 2
|
|
|
|
assert span.text == 'a sentence'
|
|
|
|
assert span.root.text == 'sentence'
|
|
|
|
assert span.root.head.text == 'is'
|
2016-01-16 18:19:09 +03:00
|
|
|
|
2017-03-11 03:50:02 +03:00
|
|
|
def test_spans_string_fn(doc):
|
|
|
|
span = doc[0:4]
|
|
|
|
assert len(span) == 4
|
|
|
|
assert span.text == 'This is a sentence'
|
|
|
|
assert span.upper_ == 'THIS IS A SENTENCE'
|
|
|
|
assert span.lower_ == 'this is a sentence'
|
2016-01-16 18:19:09 +03:00
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_root2(en_tokenizer):
|
|
|
|
text = "through North and South Carolina"
|
|
|
|
heads = [0, 3, -1, -2, -4]
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
doc = get_doc(tokens.vocab, [t.text for t in tokens], heads=heads)
|
|
|
|
assert doc[-2:].root.text == 'Carolina'
|
2016-05-06 01:17:38 +03:00
|
|
|
|
|
|
|
|
2017-01-12 17:30:49 +03:00
|
|
|
def test_spans_span_sent(doc):
|
|
|
|
"""Test span.sent property"""
|
2016-05-06 01:17:38 +03:00
|
|
|
assert len(list(doc.sents))
|
2017-01-12 17:30:49 +03:00
|
|
|
assert doc[:2].sent.root.text == 'is'
|
|
|
|
assert doc[:2].sent.text == 'This is a sentence .'
|
|
|
|
assert doc[6:7].sent.root.left_edge.text == 'This'
|
|
|
|
|
|
|
|
|
|
|
|
def test_spans_default_sentiment(en_tokenizer):
|
|
|
|
"""Test span.sentiment property's default averaging behaviour"""
|
|
|
|
text = "good stuff bad stuff"
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
tokens.vocab[tokens[0].text].sentiment = 3.0
|
|
|
|
tokens.vocab[tokens[2].text].sentiment = -2.0
|
|
|
|
doc = get_doc(tokens.vocab, [t.text for t in tokens])
|
|
|
|
assert doc[:2].sentiment == 3.0 / 2
|
|
|
|
assert doc[-2:].sentiment == -2. / 2
|
|
|
|
assert doc[:-1].sentiment == (3.+-2) / 3.
|
|
|
|
|
|
|
|
|
|
|
|
def test_spans_override_sentiment(en_tokenizer):
|
|
|
|
"""Test span.sentiment property's default averaging behaviour"""
|
|
|
|
text = "good stuff bad stuff"
|
|
|
|
tokens = en_tokenizer(text)
|
|
|
|
tokens.vocab[tokens[0].text].sentiment = 3.0
|
|
|
|
tokens.vocab[tokens[2].text].sentiment = -2.0
|
|
|
|
doc = get_doc(tokens.vocab, [t.text for t in tokens])
|
2016-12-02 13:05:50 +03:00
|
|
|
doc.user_span_hooks['sentiment'] = lambda span: 10.0
|
2017-01-12 17:30:49 +03:00
|
|
|
assert doc[:2].sentiment == 10.0
|
|
|
|
assert doc[-2:].sentiment == 10.0
|
|
|
|
assert doc[:-1].sentiment == 10.0
|