spaCy/spacy/tests/doc/test_array.py

43 lines
1.4 KiB
Python
Raw Normal View History

2015-01-30 08:43:55 +03:00
# coding: utf-8
from __future__ import unicode_literals
from ...attrs import ORTH, SHAPE, POS, DEP
from ..util import get_doc
2015-01-30 08:43:55 +03:00
import pytest
2015-01-30 08:43:55 +03:00
2017-01-11 21:00:52 +03:00
def test_doc_array_attr_of_token(en_tokenizer, en_vocab):
text = "An example sentence"
tokens = en_tokenizer(text)
example = tokens.vocab["example"]
2015-01-30 08:43:55 +03:00
assert example.orth != example.shape
feats_array = tokens.to_array((ORTH, SHAPE))
assert feats_array[0][0] != feats_array[0][1]
2015-01-30 08:43:55 +03:00
assert feats_array[0][0] != feats_array[0][1]
2017-01-11 21:00:52 +03:00
def test_doc_array_tag(en_tokenizer):
text = "A nice sentence."
pos = ['DET', 'ADJ', 'NOUN', 'PUNCT']
tokens = en_tokenizer(text)
doc = get_doc(tokens.vocab, [t.text for t in tokens], pos=pos)
assert doc[0].pos != doc[1].pos != doc[2].pos != doc[3].pos
feats_array = doc.to_array((ORTH, POS))
assert feats_array[0][1] == doc[0].pos
assert feats_array[1][1] == doc[1].pos
assert feats_array[2][1] == doc[2].pos
assert feats_array[3][1] == doc[3].pos
2017-01-11 21:00:52 +03:00
def test_doc_array_dep(en_tokenizer):
text = "A nice sentence."
deps = ['det', 'amod', 'ROOT', 'punct']
tokens = en_tokenizer(text)
doc = get_doc(tokens.vocab, [t.text for t in tokens], deps=deps)
feats_array = doc.to_array((ORTH, DEP))
assert feats_array[0][1] == doc[0].dep
assert feats_array[1][1] == doc[1].dep
assert feats_array[2][1] == doc[2].dep
assert feats_array[3][1] == doc[3].dep