mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-31 16:07:41 +03:00 
			
		
		
		
	
		
			
				
	
	
		
			77 lines
		
	
	
		
			3.0 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
			
		
		
	
	
			77 lines
		
	
	
		
			3.0 KiB
		
	
	
	
		
			Plaintext
		
	
	
	
	
	
| //- 💫 DOCS > USAGE > MODELS > LANGUAGE SUPPORT
 | |
| 
 | |
| p spaCy currently provides models for the following languages:
 | |
| 
 | |
| +table(["Language", "Code", "Language data", "Models"])
 | |
|     for models, code in MODELS
 | |
|         - var count = Object.keys(models).length
 | |
|         +row
 | |
|             +cell=LANGUAGES[code]
 | |
|             +cell #[code=code]
 | |
|             +cell
 | |
|                 +src(gh("spaCy", "spacy/lang/" + code)) #[code lang/#{code}]
 | |
|             +cell
 | |
|                 +a("/models/" + code) #{count} #{(count == 1) ? "model" : "models"}
 | |
| 
 | |
| +h(3, "alpha-support") Alpha tokenization support
 | |
| 
 | |
| p
 | |
|     |  Work has started on the following languages. You can help by
 | |
|     |  #[+a("/usage/adding-languages#language-data") improving the existing language data]
 | |
|     |  and extending the tokenization patterns.
 | |
| 
 | |
| +aside("Usage note")
 | |
|     |  Note that the alpha languages don't yet come with a language model. In
 | |
|     |  order to use them, you have to import them directly, or use
 | |
|     |  #[+api("spacy#blank") #[code spacy.blank]]:
 | |
| 
 | |
|     +code.o-no-block.
 | |
|         from spacy.lang.fi import Finnish
 | |
|         nlp = Finnish()  # use directly
 | |
|         nlp = spacy.blank('fi')  # blank instance
 | |
| 
 | |
| +table(["Language", "Code", "Language data"])
 | |
|     for lang, code in LANGUAGES
 | |
|         if !Object.keys(MODELS).includes(code)
 | |
|             +row
 | |
|                 +cell #{LANGUAGES[code]}
 | |
|                 +cell #[code=code]
 | |
|                 +cell
 | |
|                     +src(gh("spaCy", "spacy/lang/" + code)) #[code lang/#{code}]
 | |
| 
 | |
| +infobox("Dependencies")
 | |
|     .o-block-small Some language tokenizers require external dependencies.
 | |
| 
 | |
|     +list.o-no-block
 | |
|         +item #[strong Chinese]: #[+a("https://github.com/fxsjy/jieba") Jieba]
 | |
|         +item #[strong Japanese]: #[+a("https://github.com/mocobeta/janome") Janome]
 | |
|         +item #[strong Thai]: #[+a("https://github.com/wannaphongcom/pythainlp") pythainlp]
 | |
|         +item #[strong Vietnamese]: #[+a("https://github.com/trungtv/pyvi") Pyvi]
 | |
|         +item #[strong Russian]: #[+a("https://github.com/kmike/pymorphy2") pymorphy2]
 | |
| 
 | |
| +h(3, "multi-language") Multi-language support
 | |
|     +tag-new(2)
 | |
| 
 | |
| p
 | |
|     |  As of v2.0, spaCy supports models trained on more than one language. This
 | |
|     |  is especially useful for named entity recognition. The language ID used
 | |
|     |  for multi-language or language-neutral models is #[code xx]. The
 | |
|     |  language class, a generic subclass containing only the base language data,
 | |
|     |  can be found in #[+src(gh("spaCy", "spacy/lang/xx")) #[code lang/xx]].
 | |
| 
 | |
| p
 | |
|     |  To load your model with the neutral, multi-language class, simply set
 | |
|     |  #[code "language": "xx"] in your
 | |
|     |  #[+a("/usage/training#models-generating") model package]'s
 | |
|     |  meta.json. You can also import the class directly, or call
 | |
|     |  #[+api("util#get_lang_class") #[code util.get_lang_class()]] for
 | |
|     |  lazy-loading.
 | |
| 
 | |
| +code("Standard import").
 | |
|     from spacy.lang.xx import MultiLanguage
 | |
|     nlp = MultiLanguage()
 | |
| 
 | |
| +code("With lazy-loading").
 | |
|     from spacy.util import get_lang_class
 | |
|     nlp = get_lang_class('xx')
 |