mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 09:57:26 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			61 lines
		
	
	
		
			2.2 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
			
		
		
	
	
			61 lines
		
	
	
		
			2.2 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
//- 💫 DOCS > API > LANGUAGE MODELS
 | 
						|
 | 
						|
include ../../_includes/_mixins
 | 
						|
 | 
						|
p spaCy currently supports the following languages and capabilities:
 | 
						|
 | 
						|
+aside-code("Download language models", "bash").
 | 
						|
    python -m spacy download en
 | 
						|
    python -m spacy download de
 | 
						|
    python -m spacy download fr
 | 
						|
 | 
						|
+table([ "Language", "Token", "SBD", "Lemma", "POS", "NER", "Dep", "Vector", "Sentiment"])
 | 
						|
    +row
 | 
						|
        +cell English #[code en]
 | 
						|
        each icon in [ "pro", "pro", "pro", "pro", "pro", "pro", "pro", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
    +row
 | 
						|
        +cell German #[code de]
 | 
						|
        each icon in [ "pro", "pro", "con", "pro", "pro", "pro", "pro", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
    +row
 | 
						|
        +cell French #[code fr]
 | 
						|
        each icon in [ "pro", "pro", "con", "pro", "con", "pro", "pro", "con" ]
 | 
						|
            +cell.u-text-center #[+procon(icon)]
 | 
						|
 | 
						|
+h(2, "available") Available models
 | 
						|
 | 
						|
include ../usage/_models-list
 | 
						|
 | 
						|
+h(2, "alpha-support") Alpha tokenization support
 | 
						|
 | 
						|
p
 | 
						|
    |  Work has started on the following languages. You can help by
 | 
						|
    |  #[+a("/docs/usage/adding-languages#language-data") improving the existing language data]
 | 
						|
    |  and extending the tokenization patterns.
 | 
						|
 | 
						|
+aside("Usage note")
 | 
						|
    |  Note that the alpha languages don't yet come with a language model. In
 | 
						|
    |  order to use them, you have to import them directly:
 | 
						|
 | 
						|
    +code.o-no-block.
 | 
						|
        from spacy.lang.fi import Finnish
 | 
						|
        nlp = Finnish()
 | 
						|
        doc = nlp(u'Ilmatyynyalukseni on täynnä ankeriaita')
 | 
						|
 | 
						|
+infobox("Dependencies")
 | 
						|
    |  Some language tokenizers require external dependencies. To use #[strong Chinese],
 | 
						|
    |  you need to have #[+a("https://github.com/fxsjy/jieba") Jieba] installed.
 | 
						|
    |  The #[strong Japanese] tokenizer requires
 | 
						|
    |  #[+a("https://github.com/mocobeta/janome") Janome].
 | 
						|
 | 
						|
+table([ "Language", "Code", "Source" ])
 | 
						|
    each language, code in { es: "Spanish", it: "Italian", pt: "Portuguese", nl: "Dutch", sv: "Swedish", fi: "Finnish", nb: "Norwegian Bokmål", da: "Danish", hu: "Hungarian", pl: "Polish", bn: "Bengali", he: "Hebrew", zh: "Chinese", ja: "Japanese" }
 | 
						|
        +row
 | 
						|
            +cell #{language}
 | 
						|
            +cell #[code=code]
 | 
						|
            +cell
 | 
						|
                +src(gh("spaCy", "spacy/lang/" + code)) lang/#{code}
 |