mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-31 16:07:41 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			20 lines
		
	
	
		
			365 B
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			20 lines
		
	
	
		
			365 B
		
	
	
	
		
			Python
		
	
	
	
	
	
| # coding: utf8
 | ||
| from __future__ import unicode_literals
 | ||
| 
 | ||
| import pytest
 | ||
| 
 | ||
| 
 | ||
| @pytest.mark.parametrize(
 | ||
|     "word",
 | ||
|     [
 | ||
|         "don't",
 | ||
|         "don’t",
 | ||
|         "I'd",
 | ||
|         "I’d",
 | ||
|     ],
 | ||
| )
 | ||
| def test_issue3521(en_tokenizer, word):
 | ||
|     tok = en_tokenizer(word)[1]
 | ||
|     # 'not' and 'would' should be stopwords, also in their abbreviated forms
 | ||
|     assert tok.is_stop
 |