mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-11 20:28:20 +03:00
126 lines
6.0 KiB
Python
126 lines
6.0 KiB
Python
import pytest
|
|
from spacy.tokens import Doc
|
|
|
|
|
|
@pytest.fixture
|
|
def words():
|
|
# fmt: off
|
|
return [
|
|
"\n", "It", "was", "a", "bright", "cold", "day", "in", "April", ",",
|
|
"and", "the", "clocks", "were", "striking", "thirteen", ".", "\n",
|
|
"Winston", "Smith", ",", "his", "chin", "nuzzled", "into", "his",
|
|
"breast", "in", "an", "effort", "to", "escape", "the", "\n", "vile",
|
|
"wind", ",", "slipped", "quickly", "through", "the", "glass", "doors",
|
|
"of", "Victory", "Mansions", ",", "\n", "though", "not", "quickly",
|
|
"enough", "to", "prevent", "a", "swirl", "of", "gritty", "dust",
|
|
"from", "entering", "\n", "along", "with", "him", ".", "\n\n", "The",
|
|
"hallway", "smelt", "of", "boiled", "cabbage", "and", "old", "rag",
|
|
"mats", ".", "At", "one", "end", "of", "it", "a", "\n", "coloured",
|
|
"poster", ",", "too", "large", "for", "indoor", "display", ",", "had",
|
|
"been", "tacked", "to", "the", "wall", ".", "\n", "It", "depicted",
|
|
"simply", "an", "enormous", "face", ",", "more", "than", "a", "metre",
|
|
"wide", ":", "the", "face", "of", "a", "\n", "man", "of", "about",
|
|
"forty", "-", "five", ",", "with", "a", "heavy", "black", "moustache",
|
|
"and", "ruggedly", "handsome", "\n", "features", ".", "Winston", "made",
|
|
"for", "the", "stairs", ".", "It", "was", "no", "use", "trying", "the",
|
|
"lift", ".", "Even", "at", "\n", "the", "best", "of", "times", "it",
|
|
"was", "seldom", "working", ",", "and", "at", "present", "the",
|
|
"electric", "current", "\n", "was", "cut", "off", "during", "daylight",
|
|
"hours", ".", "It", "was", "part", "of", "the", "economy", "drive",
|
|
"in", "\n", "preparation", "for", "Hate", "Week", ".", "The", "flat",
|
|
"was", "seven", "flights", "up", ",", "and", "Winston", ",", "who",
|
|
"\n", "was", "thirty", "-", "nine", "and", "had", "a", "varicose",
|
|
"ulcer", "above", "his", "right", "ankle", ",", "went", "slowly", ",",
|
|
"\n", "resting", "several", "times", "on", "the", "way", ".", "On",
|
|
"each", "landing", ",", "opposite", "the", "lift", "-", "shaft", ",",
|
|
"\n", "the", "poster", "with", "the", "enormous", "face", "gazed",
|
|
"from", "the", "wall", ".", "It", "was", "one", "of", "those", "\n",
|
|
"pictures", "which", "are", "so", "contrived", "that", "the", "eyes",
|
|
"follow", "you", "about", "when", "you", "move", ".", "\n", "BIG",
|
|
"BROTHER", "IS", "WATCHING", "YOU", ",", "the", "caption", "beneath",
|
|
"it", "ran", ".", "\n", ]
|
|
# fmt: on
|
|
|
|
|
|
@pytest.fixture
|
|
def heads():
|
|
# fmt: off
|
|
return [
|
|
1, 2, 2, 6, 6, 6, 2, 6, 7, 2, 2, 12, 14, 14, 2, 14, 14, 16, 19, 23, 23,
|
|
22, 23, 23, 23, 26, 24, 23, 29, 27, 31, 29, 35, 32, 35, 31, 23, 23, 37,
|
|
37, 42, 42, 39, 42, 45, 43, 37, 46, 37, 50, 51, 37, 53, 51, 55, 53, 55,
|
|
58, 56, 53, 59, 60, 60, 62, 63, 23, 65, 68, 69, 69, 69, 72, 70, 72, 76,
|
|
76, 72, 69, 96, 80, 78, 80, 81, 86, 83, 86, 96, 96, 89, 96, 89, 92, 90,
|
|
96, 96, 96, 96, 96, 99, 97, 96, 100, 103, 103, 103, 107, 107, 103, 107,
|
|
111, 111, 112, 113, 107, 103, 116, 136, 116, 120, 118, 117, 120, 125,
|
|
125, 125, 121, 116, 116, 131, 131, 131, 127, 131, 134, 131, 134, 136,
|
|
136, 139, 139, 139, 142, 140, 139, 145, 145, 147, 145, 147, 150, 148,
|
|
145, 153, 162, 153, 156, 162, 156, 157, 162, 162, 162, 162, 162, 162,
|
|
172, 165, 169, 169, 172, 169, 172, 162, 172, 172, 176, 174, 172, 179,
|
|
179, 179, 180, 183, 181, 179, 184, 185, 185, 187, 190, 188, 179, 193,
|
|
194, 194, 196, 194, 196, 194, 194, 218, 200, 204, 202, 200, 207, 207,
|
|
204, 204, 204, 212, 212, 209, 212, 216, 216, 213, 200, 194, 218, 218,
|
|
220, 218, 224, 222, 222, 227, 225, 218, 246, 231, 229, 246, 246, 237,
|
|
237, 237, 233, 246, 238, 241, 246, 241, 245, 245, 242, 246, 246, 249,
|
|
247, 246, 252, 252, 252, 253, 257, 255, 254, 259, 257, 261, 259, 265,
|
|
264, 265, 261, 265, 265, 270, 270, 267, 252, 271, 274, 275, 275, 276,
|
|
283, 283, 280, 283, 280, 281, 283, 283, 284]
|
|
# fmt: on
|
|
|
|
|
|
def test_parser_parse_navigate_consistency(en_vocab, words, heads):
|
|
doc = Doc(en_vocab, words=words, heads=heads)
|
|
for head in doc:
|
|
for child in head.lefts:
|
|
assert child.head == head
|
|
for child in head.rights:
|
|
assert child.head == head
|
|
|
|
|
|
def test_parser_parse_navigate_child_consistency(en_vocab, words, heads):
|
|
doc = Doc(en_vocab, words=words, heads=heads, deps=["dep"] * len(heads))
|
|
lefts = {}
|
|
rights = {}
|
|
for head in doc:
|
|
assert head.i not in lefts
|
|
lefts[head.i] = set()
|
|
for left in head.lefts:
|
|
lefts[head.i].add(left.i)
|
|
assert head.i not in rights
|
|
rights[head.i] = set()
|
|
for right in head.rights:
|
|
rights[head.i].add(right.i)
|
|
for head in doc:
|
|
assert head.n_rights == len(rights[head.i])
|
|
assert head.n_lefts == len(lefts[head.i])
|
|
for child in doc:
|
|
if child.i < child.head.i:
|
|
assert child.i in lefts[child.head.i]
|
|
assert child.i not in rights[child.head.i]
|
|
lefts[child.head.i].remove(child.i)
|
|
elif child.i > child.head.i:
|
|
assert child.i in rights[child.head.i]
|
|
assert child.i not in lefts[child.head.i]
|
|
rights[child.head.i].remove(child.i)
|
|
for head_index, children in lefts.items():
|
|
assert not children
|
|
for head_index, children in rights.items():
|
|
assert not children
|
|
|
|
|
|
def test_parser_parse_navigate_edges(en_vocab, words, heads):
|
|
doc = Doc(en_vocab, words=words, heads=heads)
|
|
for token in doc:
|
|
subtree = list(token.subtree)
|
|
debug = "\t".join((token.text, token.left_edge.text, subtree[0].text))
|
|
assert token.left_edge == subtree[0], debug
|
|
debug = "\t".join(
|
|
(
|
|
token.text,
|
|
token.right_edge.text,
|
|
subtree[-1].text,
|
|
token.right_edge.head.text,
|
|
)
|
|
)
|
|
assert token.right_edge == subtree[-1], debug
|