mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-31 16:07:41 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			32 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
			
		
		
	
	
			32 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
| //- 💫 DOCS > API > LANGUAGE MODELS
 | |
| 
 | |
| include ../../_includes/_mixins
 | |
| 
 | |
| p You can download data packs that add the following capabilities to spaCy.
 | |
| 
 | |
| +aside-code("Download language models", "bash").
 | |
|     python -m spacy.en.download all
 | |
|     python -m spacy.de.download all
 | |
| 
 | |
| +table([ "Language", "Token", "SBD", "Lemma", "POS", "NER", "Dep", "Vector", "Sentiment"])
 | |
|     +row
 | |
|         +cell English #[code en]
 | |
|         each icon in [ "pro", "pro", "pro", "pro", "pro", "pro", "pro", "con" ]
 | |
|             +cell.u-text-center #[+procon(icon)]
 | |
| 
 | |
|     +row
 | |
|         +cell German #[code de]
 | |
|         each icon in [ "pro", "pro", "con", "pro", "pro", "pro", "pro", "con" ]
 | |
|             +cell.u-text-center #[+procon(icon)]
 | |
| 
 | |
|     +row
 | |
|         +cell Chinese #[code zh]
 | |
|         each icon in [ "pro", "con", "con", "con", "con", "con", "con", "con" ]
 | |
|             +cell.u-text-center #[+procon(icon)]
 | |
| 
 | |
| p
 | |
|     |  Chinese tokenization requires the
 | |
|     |  #[+a("https://github.com/fxsjy/jieba") Jieba] library. Statistical
 | |
|     |  models are coming soon. Tokenizers for Spanish, French, Italian and
 | |
|     |  Portuguese are now under development. 
 |