From 571b6eda88bb72078b88b9a600455cb8ed3ab622 Mon Sep 17 00:00:00 2001 From: Matthew Honnibal Date: Sun, 6 Sep 2015 05:40:10 +0200 Subject: [PATCH] * Upd tests --- tests/parser/test_initial_actions_parse.py | 5 ++++- tests/test_matcher.py | 12 ++++++------ 2 files changed, 10 insertions(+), 7 deletions(-) diff --git a/tests/parser/test_initial_actions_parse.py b/tests/parser/test_initial_actions_parse.py index c1603cd93..9f570d8be 100644 --- a/tests/parser/test_initial_actions_parse.py +++ b/tests/parser/test_initial_actions_parse.py @@ -4,7 +4,10 @@ import pytest def test_initial(EN): doc = EN.tokenizer(u'I ate the pizza with anchovies.') EN.tagger(doc) - next_actions = EN.parser.partial(doc, ['L-nsubj', 'S', 'L-det']) + with EN.parser.step_through(doc) as stepwise: + stepwise.transition('L-nsubj') + stepwise.transition('S') + stepwise.transition('L-det') assert doc[0].head.i == 1 assert doc[1].head.i == 1 assert doc[2].head.i == 3 diff --git a/tests/test_matcher.py b/tests/test_matcher.py index 0014e1110..1b748cb53 100644 --- a/tests/test_matcher.py +++ b/tests/test_matcher.py @@ -3,7 +3,7 @@ import pytest from spacy.strings import StringStore from spacy.matcher import * -from spacy.attrs import ORTH +from spacy.attrs import LOWER from spacy.tokens.doc import Doc from spacy.vocab import Vocab @@ -13,7 +13,7 @@ def matcher(EN): patterns = { 'Javascript': ['PRODUCT', {}, [[{'ORTH': 'JavaScript'}]]], 'GoogleNow': ['PRODUCT', {}, [[{'ORTH': 'Google'}, {'ORTH': 'Now'}]]], - 'Java': ['PRODUCT', {}, [[{'ORTH': 'Java'}]]], + 'Java': ['PRODUCT', {}, [[{'LOWER': 'java'}]]], } return Matcher(EN.vocab, patterns) @@ -33,7 +33,7 @@ def test_match_start(matcher, EN): def test_match_end(matcher, EN): - tokens = EN('I like Java') + tokens = EN('I like java') assert matcher(tokens) == [(EN.vocab.strings['PRODUCT'], 2, 3)] @@ -43,17 +43,17 @@ def test_match_middle(matcher, EN): def test_match_multi(matcher, EN): - tokens = EN('I like Google Now and Java best') + tokens = EN('I like Google Now and java best') assert matcher(tokens) == [(EN.vocab.strings['PRODUCT'], 2, 4), (EN.vocab.strings['PRODUCT'], 5, 6)] def test_match_preserved(matcher, EN): - doc = EN.tokenizer('I like Java') + doc = EN.tokenizer('I like java') EN.tagger(doc) EN.entity(doc) assert len(doc.ents) == 0 - doc = EN.tokenizer('I like Java') + doc = EN.tokenizer('I like java') matcher(doc) assert len(doc.ents) == 1 EN.tagger(doc)