mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-26 18:06:29 +03:00
Remove old unused tests and conftest files
This commit is contained in:
parent
4d0bfebcd9
commit
f8803808ce
1
setup.py
1
setup.py
|
@ -37,7 +37,6 @@ PACKAGES = [
|
||||||
'spacy.munge',
|
'spacy.munge',
|
||||||
'spacy.tests',
|
'spacy.tests',
|
||||||
'spacy.tests.matcher',
|
'spacy.tests.matcher',
|
||||||
'spacy.tests.morphology',
|
|
||||||
'spacy.tests.munge',
|
'spacy.tests.munge',
|
||||||
'spacy.tests.parser',
|
'spacy.tests.parser',
|
||||||
'spacy.tests.serialize',
|
'spacy.tests.serialize',
|
||||||
|
|
|
@ -1,17 +0,0 @@
|
||||||
import pytest
|
|
||||||
|
|
||||||
import pickle
|
|
||||||
import io
|
|
||||||
|
|
||||||
|
|
||||||
from spacy.morphology import Morphology
|
|
||||||
from spacy.lemmatizer import Lemmatizer
|
|
||||||
from spacy.strings import StringStore
|
|
||||||
|
|
||||||
@pytest.mark.xfail
|
|
||||||
def test_pickle():
|
|
||||||
morphology = Morphology(StringStore(), {}, Lemmatizer({}, {}, {}))
|
|
||||||
|
|
||||||
file_ = io.BytesIO()
|
|
||||||
pickle.dump(morphology, file_)
|
|
||||||
|
|
|
@ -92,22 +92,3 @@ def test_phrase_matcher():
|
||||||
matcher = PhraseMatcher(vocab, [Doc(vocab, words='Google Now'.split())])
|
matcher = PhraseMatcher(vocab, [Doc(vocab, words='Google Now'.split())])
|
||||||
doc = Doc(vocab, words=['I', 'like', 'Google', 'Now', 'best'])
|
doc = Doc(vocab, words=['I', 'like', 'Google', 'Now', 'best'])
|
||||||
assert len(matcher(doc)) == 1
|
assert len(matcher(doc)) == 1
|
||||||
|
|
||||||
|
|
||||||
#@pytest.mark.models
|
|
||||||
#def test_match_preserved(EN):
|
|
||||||
# patterns = {
|
|
||||||
# 'JS': ['PRODUCT', {}, [[{'ORTH': 'JavaScript'}]]],
|
|
||||||
# 'GoogleNow': ['PRODUCT', {}, [[{'ORTH': 'Google'}, {'ORTH': 'Now'}]]],
|
|
||||||
# 'Java': ['PRODUCT', {}, [[{'LOWER': 'java'}]]],
|
|
||||||
# }
|
|
||||||
# matcher = Matcher(EN.vocab, patterns)
|
|
||||||
# doc = EN.tokenizer('I like java.')
|
|
||||||
# EN.tagger(doc)
|
|
||||||
# assert len(doc.ents) == 0
|
|
||||||
# doc = EN.tokenizer('I like java.')
|
|
||||||
# doc.ents += tuple(matcher(doc))
|
|
||||||
# assert len(doc.ents) == 1
|
|
||||||
# EN.tagger(doc)
|
|
||||||
# EN.entity(doc)
|
|
||||||
# assert len(doc.ents) == 1
|
|
||||||
|
|
|
@ -1,6 +0,0 @@
|
||||||
import pytest
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture(scope="session")
|
|
||||||
def en_vocab(EN):
|
|
||||||
return EN.vocab
|
|
Loading…
Reference in New Issue
Block a user