mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-30 23:47:31 +03:00 
			
		
		
		
	Fix Japanese tokenizer
JapaneseTokenizer now returns a Doc, not individual words
This commit is contained in:
		
							parent
							
								
									5ae0b8613a
								
							
						
					
					
						commit
						66f8f9d4a0
					
				|  | @ -33,8 +33,7 @@ class Japanese(Language): | |||
|     Defaults = JapaneseDefaults | ||||
| 
 | ||||
|     def make_doc(self, text): | ||||
|         words = self.tokenizer(text) | ||||
|         return Doc(self.vocab, words=words, spaces=[False]*len(words)) | ||||
|         return self.tokenizer(text) | ||||
| 
 | ||||
| 
 | ||||
| __all__ = ['Japanese'] | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue
	
	Block a user