mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 09:57:26 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			48 lines
		
	
	
		
			1.6 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			48 lines
		
	
	
		
			1.6 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# coding: utf-8
 | 
						|
from __future__ import unicode_literals
 | 
						|
 | 
						|
import sys
 | 
						|
import pytest
 | 
						|
 | 
						|
 | 
						|
def test_tokenizer_handles_emoticons(tokenizer):
 | 
						|
    # Tweebo challenge (CMU)
 | 
						|
    text = """:o :/ :'( >:o (: :) >.< XD -__- o.O ;D :-) @_@ :P 8D :1 >:( :D =| ") :> ...."""
 | 
						|
    tokens = tokenizer(text)
 | 
						|
    assert tokens[0].text == ":o"
 | 
						|
    assert tokens[1].text == ":/"
 | 
						|
    assert tokens[2].text == ":'("
 | 
						|
    assert tokens[3].text == ">:o"
 | 
						|
    assert tokens[4].text == "(:"
 | 
						|
    assert tokens[5].text == ":)"
 | 
						|
    assert tokens[6].text == ">.<"
 | 
						|
    assert tokens[7].text == "XD"
 | 
						|
    assert tokens[8].text == "-__-"
 | 
						|
    assert tokens[9].text == "o.O"
 | 
						|
    assert tokens[10].text == ";D"
 | 
						|
    assert tokens[11].text == ":-)"
 | 
						|
    assert tokens[12].text == "@_@"
 | 
						|
    assert tokens[13].text == ":P"
 | 
						|
    assert tokens[14].text == "8D"
 | 
						|
    assert tokens[15].text == ":1"
 | 
						|
    assert tokens[16].text == ">:("
 | 
						|
    assert tokens[17].text == ":D"
 | 
						|
    assert tokens[18].text == "=|"
 | 
						|
    assert tokens[19].text == '")'
 | 
						|
    assert tokens[20].text == ':>'
 | 
						|
    assert tokens[21].text == '....'
 | 
						|
 | 
						|
 | 
						|
@pytest.mark.parametrize('text,length', [("example:)", 3), ("108)", 2), ("XDN", 1)])
 | 
						|
def test_tokenizer_excludes_false_pos_emoticons(tokenizer, text, length):
 | 
						|
    tokens = tokenizer(text)
 | 
						|
    assert len(tokens) == length
 | 
						|
 | 
						|
@pytest.mark.parametrize('text,length', [('can you still dunk?🍕🍔😵LOL', 8),
 | 
						|
                                         ('i💙you', 3), ('🤘🤘yay!', 4)])
 | 
						|
def test_tokenizer_handles_emoji(tokenizer, text, length):
 | 
						|
    # These break on narrow unicode builds, e.g. Windows
 | 
						|
    if sys.maxunicode >= 1114111:
 | 
						|
        tokens = tokenizer(text)
 | 
						|
        assert len(tokens) == length
 |