mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-31 16:07:41 +03:00 
			
		
		
		
	* Upd tests
This commit is contained in:
		
							parent
							
								
									20c2db08b7
								
							
						
					
					
						commit
						3a5299faec
					
				|  | @ -1,4 +1,6 @@ | ||||||
|  | import pytest | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test_simple_types(EN): | def test_simple_types(EN): | ||||||
|     tokens = EN(u'Mr. Best flew to New York on Saturday morning.') |     tokens = EN(u'Mr. Best flew to New York on Saturday morning.') | ||||||
|     ents = list(tokens.ents) |     ents = list(tokens.ents) | ||||||
|  |  | ||||||
|  | @ -96,6 +96,7 @@ def test_rosetta(): | ||||||
|     assert my_exp_len == py_exp_len |     assert my_exp_len == py_exp_len | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.slow | ||||||
| def test_vocab(EN): | def test_vocab(EN): | ||||||
|     codec = HuffmanCodec([(w.orth, numpy.exp(w.prob)) for w in EN.vocab]) |     codec = HuffmanCodec([(w.orth, numpy.exp(w.prob)) for w in EN.vocab]) | ||||||
|     expected_length = 0 |     expected_length = 0 | ||||||
|  | @ -105,6 +106,7 @@ def test_vocab(EN): | ||||||
|     assert 8 < expected_length < 15 |     assert 8 < expected_length < 15 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.slow | ||||||
| def test_freqs(): | def test_freqs(): | ||||||
|     freqs = [] |     freqs = [] | ||||||
|     words = [] |     words = [] | ||||||
|  |  | ||||||
|  | @ -120,5 +120,3 @@ def test_packer_annotated(tokenizer): | ||||||
|     assert [t.tag_ for t in result] == ['DT', 'NN', 'VBD'] |     assert [t.tag_ for t in result] == ['DT', 'NN', 'VBD'] | ||||||
|     assert [t.dep_ for t in result] == ['det', 'nsubj', 'ROOT'] |     assert [t.dep_ for t in result] == ['det', 'nsubj', 'ROOT'] | ||||||
|     assert [(t.head.i - t.i) for t in result] == [1, 1, 0] |     assert [(t.head.i - t.i) for t in result] == [1, 1, 0] | ||||||
| 
 |  | ||||||
| 
 |  | ||||||
|  |  | ||||||
|  | @ -1,6 +1,8 @@ | ||||||
| from __future__ import unicode_literals | from __future__ import unicode_literals | ||||||
|  | import pytest | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test_merge_tokens(EN): | def test_merge_tokens(EN): | ||||||
|     tokens = EN(u'Los Angeles start.') |     tokens = EN(u'Los Angeles start.') | ||||||
|     assert len(tokens) == 4 |     assert len(tokens) == 4 | ||||||
|  | @ -12,6 +14,7 @@ def test_merge_tokens(EN): | ||||||
|     assert tokens[0].head.orth_ == 'start' |     assert tokens[0].head.orth_ == 'start' | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test_merge_heads(EN): | def test_merge_heads(EN): | ||||||
|     tokens = EN(u'I found a pilates class near work.') |     tokens = EN(u'I found a pilates class near work.') | ||||||
|     assert len(tokens) == 8 |     assert len(tokens) == 8 | ||||||
|  |  | ||||||
|  | @ -1,8 +1,9 @@ | ||||||
| # -*- coding: utf-8 -*- | # -*- coding: utf-8 -*- | ||||||
| """Sphinx doctest is just too hard. Manually paste doctest examples here""" | """Sphinx doctest is just too hard. Manually paste doctest examples here""" | ||||||
| from spacy.en.attrs import IS_LOWER | from spacy.en.attrs import IS_LOWER | ||||||
|  | import pytest | ||||||
| 
 | 
 | ||||||
| 
 | @pytest.mark.models | ||||||
| def test_1(): | def test_1(): | ||||||
|     import spacy.en |     import spacy.en | ||||||
|     from spacy.parts_of_speech import ADV |     from spacy.parts_of_speech import ADV | ||||||
|  | @ -21,6 +22,7 @@ def test_1(): | ||||||
|     assert o == -11.07155704498291 |     assert o == -11.07155704498291 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test2(): | def test2(): | ||||||
|     import spacy.en |     import spacy.en | ||||||
|     from spacy.parts_of_speech import ADV |     from spacy.parts_of_speech import ADV | ||||||
|  | @ -41,6 +43,7 @@ def test2(): | ||||||
|     -11.07155704498291 |     -11.07155704498291 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test3(): | def test3(): | ||||||
|     import spacy.en |     import spacy.en | ||||||
|     from spacy.parts_of_speech import ADV |     from spacy.parts_of_speech import ADV | ||||||
|  |  | ||||||
|  | @ -7,6 +7,7 @@ from spacy.en.attrs import IS_STOP | ||||||
| import pytest | import pytest | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.models | ||||||
| def test_strings(EN): | def test_strings(EN): | ||||||
|     tokens = EN(u'Give it back! He pleaded.') |     tokens = EN(u'Give it back! He pleaded.') | ||||||
|     token = tokens[0] |     token = tokens[0] | ||||||
|  |  | ||||||
|  | @ -5,7 +5,7 @@ from spacy.tokens import Doc | ||||||
| import pytest | import pytest | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| def test_getitem(EN): | def mest_getitem(EN): | ||||||
|     tokens = EN(u'Give it back! He pleaded.') |     tokens = EN(u'Give it back! He pleaded.') | ||||||
|     assert tokens[0].orth_ == 'Give' |     assert tokens[0].orth_ == 'Give' | ||||||
|     assert tokens[-1].orth_ == '.' |     assert tokens[-1].orth_ == '.' | ||||||
|  | @ -13,10 +13,19 @@ def test_getitem(EN): | ||||||
|         tokens[len(tokens)] |         tokens[len(tokens)] | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
| def test_serialize(EN): | def mest_serialize(EN): | ||||||
|     tokens = EN(u'Give it back! He pleaded.') |     tokens = EN(u'Give it back! He pleaded.') | ||||||
|     packed = tokens.serialize() |     packed = tokens.to_bytes() | ||||||
|     new_tokens = Doc.deserialize(EN.vocab, packed) |     new_tokens = Doc(EN.vocab).from_bytes(packed) | ||||||
|  |     assert tokens.string == new_tokens.string | ||||||
|  |     assert [t.orth_ for t in tokens] == [t.orth_ for t in new_tokens] | ||||||
|  |     assert [t.orth for t in tokens] == [t.orth for t in new_tokens] | ||||||
|  | 
 | ||||||
|  | 
 | ||||||
|  | def test_serialize_whitespace(EN): | ||||||
|  |     tokens = EN(u' Give it back! He pleaded. ') | ||||||
|  |     packed = tokens.to_bytes() | ||||||
|  |     new_tokens = Doc(EN.vocab).from_bytes(packed) | ||||||
|     assert tokens.string == new_tokens.string |     assert tokens.string == new_tokens.string | ||||||
|     assert [t.orth_ for t in tokens] == [t.orth_ for t in new_tokens] |     assert [t.orth_ for t in tokens] == [t.orth_ for t in new_tokens] | ||||||
|     assert [t.orth for t in tokens] == [t.orth for t in new_tokens] |     assert [t.orth for t in tokens] == [t.orth for t in new_tokens] | ||||||
|  |  | ||||||
|  | @ -4,13 +4,14 @@ from spacy.en import English | ||||||
| 
 | 
 | ||||||
| import pytest | import pytest | ||||||
| 
 | 
 | ||||||
| 
 | @pytest.mark.vectors | ||||||
| def test_vec(EN): | def test_vec(EN): | ||||||
|     hype = EN.vocab['hype'] |     hype = EN.vocab['hype'] | ||||||
|     assert hype.orth_ == 'hype' |     assert hype.orth_ == 'hype' | ||||||
|     assert 0.08 >= hype.repvec[0] > 0.07 |     assert 0.08 >= hype.repvec[0] > 0.07 | ||||||
| 
 | 
 | ||||||
| 
 | 
 | ||||||
|  | @pytest.mark.vectors | ||||||
| def test_capitalized(EN): | def test_capitalized(EN): | ||||||
|     hype = EN.vocab['Hype'] |     hype = EN.vocab['Hype'] | ||||||
|     assert hype.orth_ == 'Hype' |     assert hype.orth_ == 'Hype' | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue
	
	Block a user