mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 01:48:04 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			32 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
			
		
		
	
	
			32 lines
		
	
	
		
			1.1 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
//- 💫 DOCS > API > LANGUAGE MODELS
 | 
						|
 | 
						|
include ../../_includes/_mixins
 | 
						|
 | 
						|
p You can download data packs that add the following capabilities to spaCy.
 | 
						|
 | 
						|
+aside-code("Download language models", "bash").
 | 
						|
    python -m spacy.en.download all
 | 
						|
    python -m spacy.de.download all
 | 
						|
 | 
						|
+table([ "Language", "Token", "SBD", "Lemma", "POS", "NER", "Dep", "Vector", "Sentiment"])
 | 
						|
    +row
 | 
						|
        +cell English #[code en]
 | 
						|
        each icon in [ "pro", "pro", "pro", "pro", "pro", "pro", "pro", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
    +row
 | 
						|
        +cell German #[code de]
 | 
						|
        each icon in [ "pro", "pro", "con", "pro", "pro", "pro", "pro", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
    +row
 | 
						|
        +cell Chinese #[code zh]
 | 
						|
        each icon in [ "pro", "con", "con", "con", "con", "con", "con", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
p
 | 
						|
    |  Chinese tokenization requires the
 | 
						|
    |  #[+a("https://github.com/fxsjy/jieba") Jieba] library. Statistical
 | 
						|
    |  models are coming soon. Tokenizers for Spanish, French, Italian and
 | 
						|
    |  Portuguese are now under development. 
 |