mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 09:57:26 +03:00 
			
		
		
		
	* Update tests, preventing the parser from being loaded if possible
This commit is contained in:
		
							parent
							
								
									a9c70c9447
								
							
						
					
					
						commit
						89c3364041
					
				| 
						 | 
					@ -11,7 +11,7 @@ def EN():
 | 
				
			||||||
@pytest.fixture
 | 
					@pytest.fixture
 | 
				
			||||||
def tagged(EN):
 | 
					def tagged(EN):
 | 
				
			||||||
    string = u'Bananas in pyjamas are geese.'
 | 
					    string = u'Bananas in pyjamas are geese.'
 | 
				
			||||||
    tokens = EN(string, tag=True)
 | 
					    tokens = EN(string, tag=True, parse=False)
 | 
				
			||||||
    return tokens
 | 
					    return tokens
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -11,7 +11,7 @@ EN = English()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_attr_of_token():
 | 
					def test_attr_of_token():
 | 
				
			||||||
    text = u'An example sentence.'
 | 
					    text = u'An example sentence.'
 | 
				
			||||||
    tokens = EN(text)
 | 
					    tokens = EN(text, tag=True, parse=False)
 | 
				
			||||||
    example = EN.vocab[u'example']
 | 
					    example = EN.vocab[u'example']
 | 
				
			||||||
    assert example.orth != example.shape
 | 
					    assert example.orth != example.shape
 | 
				
			||||||
    feats_array = tokens.to_array((attrs.ORTH, attrs.SHAPE))
 | 
					    feats_array = tokens.to_array((attrs.ORTH, attrs.SHAPE))
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -11,7 +11,7 @@ def orths(tokens):
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_simple_two():
 | 
					def test_simple_two():
 | 
				
			||||||
    tokens = NLU('I lost money and pride.')
 | 
					    tokens = NLU('I lost money and pride.', tag=True, parse=False)
 | 
				
			||||||
    pride = tokens[4]
 | 
					    pride = tokens[4]
 | 
				
			||||||
    assert orths(pride.conjuncts) == ['money', 'pride']
 | 
					    assert orths(pride.conjuncts) == ['money', 'pride']
 | 
				
			||||||
    money = tokens[2]
 | 
					    money = tokens[2]
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3,26 +3,23 @@ import pytest
 | 
				
			||||||
 | 
					
 | 
				
			||||||
from spacy.en import English
 | 
					from spacy.en import English
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@pytest.fixture
 | 
					EN = English()
 | 
				
			||||||
def EN():
 | 
					 | 
				
			||||||
    return English()
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					def test_possess():
 | 
				
			||||||
def test_possess(EN):
 | 
					    tokens = EN("Mike's", parse=False, tag=False)
 | 
				
			||||||
    tokens = EN("Mike's", parse=False)
 | 
					 | 
				
			||||||
    assert EN.vocab.strings[tokens[0].orth] == "Mike"
 | 
					    assert EN.vocab.strings[tokens[0].orth] == "Mike"
 | 
				
			||||||
    assert EN.vocab.strings[tokens[1].orth] == "'s"
 | 
					    assert EN.vocab.strings[tokens[1].orth] == "'s"
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_apostrophe(EN):
 | 
					def test_apostrophe():
 | 
				
			||||||
    tokens = EN("schools'")
 | 
					    tokens = EN("schools'", parse=False, tag=False)
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
    assert tokens[1].orth_ == "'"
 | 
					    assert tokens[1].orth_ == "'"
 | 
				
			||||||
    assert tokens[0].orth_ == "schools"
 | 
					    assert tokens[0].orth_ == "schools"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_LL(EN):
 | 
					def test_LL():
 | 
				
			||||||
    tokens = EN("we'll", parse=False)
 | 
					    tokens = EN("we'll", parse=False)
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
    assert tokens[1].orth_ == "'ll"
 | 
					    assert tokens[1].orth_ == "'ll"
 | 
				
			||||||
| 
						 | 
					@ -30,7 +27,7 @@ def test_LL(EN):
 | 
				
			||||||
    assert tokens[0].orth_ == "we"
 | 
					    assert tokens[0].orth_ == "we"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_aint(EN):
 | 
					def test_aint():
 | 
				
			||||||
    tokens = EN("ain't", parse=False)
 | 
					    tokens = EN("ain't", parse=False)
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
    assert tokens[0].orth_ == "ai"
 | 
					    assert tokens[0].orth_ == "ai"
 | 
				
			||||||
| 
						 | 
					@ -39,7 +36,7 @@ def test_aint(EN):
 | 
				
			||||||
    assert tokens[1].lemma_ == "not"
 | 
					    assert tokens[1].lemma_ == "not"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_capitalized(EN):
 | 
					def test_capitalized():
 | 
				
			||||||
    tokens = EN("can't", parse=False)
 | 
					    tokens = EN("can't", parse=False)
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
    tokens = EN("Can't", parse=False)
 | 
					    tokens = EN("Can't", parse=False)
 | 
				
			||||||
| 
						 | 
					@ -50,7 +47,7 @@ def test_capitalized(EN):
 | 
				
			||||||
    assert tokens[0].lemma_ == "be"
 | 
					    assert tokens[0].lemma_ == "be"
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_punct(EN):
 | 
					def test_punct():
 | 
				
			||||||
    tokens = EN("We've", parse=False)
 | 
					    tokens = EN("We've", parse=False)
 | 
				
			||||||
    assert len(tokens) == 2
 | 
					    assert len(tokens) == 2
 | 
				
			||||||
    tokens = EN("``We've", parse=False)
 | 
					    tokens = EN("``We've", parse=False)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -11,7 +11,7 @@ def EN():
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_tweebo_challenge(EN):
 | 
					def test_tweebo_challenge(EN):
 | 
				
			||||||
    text = u""":o :/ :'( >:o (: :) >.< XD -__- o.O ;D :-) @_@ :P 8D :1 >:( :D =| ") :> ...."""
 | 
					    text = u""":o :/ :'( >:o (: :) >.< XD -__- o.O ;D :-) @_@ :P 8D :1 >:( :D =| ") :> ...."""
 | 
				
			||||||
    tokens = EN(text)
 | 
					    tokens = EN(text, parse=False, tag=False)
 | 
				
			||||||
    assert tokens[0].orth_ == ":o"
 | 
					    assert tokens[0].orth_ == ":o"
 | 
				
			||||||
    assert tokens[1].orth_ == ":/"
 | 
					    assert tokens[1].orth_ == ":/"
 | 
				
			||||||
    assert tokens[2].orth_ == ":'("
 | 
					    assert tokens[2].orth_ == ":'("
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -12,7 +12,7 @@ from spacy.en import English
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_period():
 | 
					def test_period():
 | 
				
			||||||
    EN = English()
 | 
					    EN = English()
 | 
				
			||||||
    tokens = EN('best.Known')
 | 
					    tokens = EN.tokenizer('best.Known')
 | 
				
			||||||
    assert len(tokens) == 3
 | 
					    assert len(tokens) == 3
 | 
				
			||||||
    tokens = EN('zombo.com')
 | 
					    tokens = EN('zombo.com')
 | 
				
			||||||
    assert len(tokens) == 1
 | 
					    assert len(tokens) == 1
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -20,7 +20,7 @@ def morph_exc():
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_load_exc(EN, morph_exc):
 | 
					def test_load_exc(EN, morph_exc):
 | 
				
			||||||
    EN.tagger.load_morph_exceptions(morph_exc)
 | 
					    EN.tagger.load_morph_exceptions(morph_exc)
 | 
				
			||||||
    tokens = EN('I like his style.', tag=True)
 | 
					    tokens = EN('I like his style.', tag=True, parse=False)
 | 
				
			||||||
    his = tokens[2]
 | 
					    his = tokens[2]
 | 
				
			||||||
    assert his.tag_ == 'PRP$'
 | 
					    assert his.tag_ == 'PRP$'
 | 
				
			||||||
    assert his.lemma_ == '-PRP-'
 | 
					    assert his.lemma_ == '-PRP-'
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -19,7 +19,7 @@ def test_close(close_puncts, EN):
 | 
				
			||||||
    word_str = 'Hello'
 | 
					    word_str = 'Hello'
 | 
				
			||||||
    for p in close_puncts:
 | 
					    for p in close_puncts:
 | 
				
			||||||
        string = word_str + p
 | 
					        string = word_str + p
 | 
				
			||||||
        tokens = EN(string)
 | 
					        tokens = EN(string, parse=False, tag=False)
 | 
				
			||||||
        assert len(tokens) == 2
 | 
					        assert len(tokens) == 2
 | 
				
			||||||
        assert tokens[1].string == p
 | 
					        assert tokens[1].string == p
 | 
				
			||||||
        assert tokens[0].string == word_str
 | 
					        assert tokens[0].string == word_str
 | 
				
			||||||
| 
						 | 
					@ -29,7 +29,7 @@ def test_two_different_close(close_puncts, EN):
 | 
				
			||||||
    word_str = 'Hello'
 | 
					    word_str = 'Hello'
 | 
				
			||||||
    for p in close_puncts:
 | 
					    for p in close_puncts:
 | 
				
			||||||
        string = word_str + p + "'"
 | 
					        string = word_str + p + "'"
 | 
				
			||||||
        tokens = EN(string)
 | 
					        tokens = EN(string, parse=False, tag=False)
 | 
				
			||||||
        assert len(tokens) == 3
 | 
					        assert len(tokens) == 3
 | 
				
			||||||
        assert tokens[0].string == word_str
 | 
					        assert tokens[0].string == word_str
 | 
				
			||||||
        assert tokens[1].string == p
 | 
					        assert tokens[1].string == p
 | 
				
			||||||
| 
						 | 
					@ -40,12 +40,12 @@ def test_three_same_close(close_puncts, EN):
 | 
				
			||||||
    word_str = 'Hello'
 | 
					    word_str = 'Hello'
 | 
				
			||||||
    for p in close_puncts:
 | 
					    for p in close_puncts:
 | 
				
			||||||
        string = word_str + p + p + p
 | 
					        string = word_str + p + p + p
 | 
				
			||||||
        tokens = EN(string)
 | 
					        tokens = EN(string, tag=False, parse=False)
 | 
				
			||||||
        assert len(tokens) == 4
 | 
					        assert len(tokens) == 4
 | 
				
			||||||
        assert tokens[0].string == word_str
 | 
					        assert tokens[0].string == word_str
 | 
				
			||||||
        assert tokens[1].string == p
 | 
					        assert tokens[1].string == p
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_double_end_quote(EN):
 | 
					def test_double_end_quote(EN):
 | 
				
			||||||
    assert len(EN("Hello''")) == 2
 | 
					    assert len(EN("Hello''", tag=False, parse=False)) == 2
 | 
				
			||||||
    assert len(EN("''")) == 1
 | 
					    assert len(EN("''", tag=False, parse=False)) == 1
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -12,7 +12,7 @@ def paired_puncts():
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@pytest.fixture
 | 
					@pytest.fixture
 | 
				
			||||||
def EN():
 | 
					def EN():
 | 
				
			||||||
    return English()
 | 
					    return English().tokenizer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_token(paired_puncts, EN):
 | 
					def test_token(paired_puncts, EN):
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -7,7 +7,7 @@ import pytest
 | 
				
			||||||
 | 
					
 | 
				
			||||||
@pytest.fixture
 | 
					@pytest.fixture
 | 
				
			||||||
def EN():
 | 
					def EN():
 | 
				
			||||||
    return English()
 | 
					    return English().tokenizer
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
def test_single_space(EN):
 | 
					def test_single_space(EN):
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in New Issue
	
	Block a user