From a5da683578db7711073f37a6fadf34def5305ea4 Mon Sep 17 00:00:00 2001 From: ines Date: Sat, 14 Oct 2017 12:52:41 +0200 Subject: [PATCH] Add Russian to alpha docs and update tokenizer dependencies --- website/models/_data.json | 1 + website/usage/_models/_languages.jade | 11 +++++++---- 2 files changed, 8 insertions(+), 4 deletions(-) diff --git a/website/models/_data.json b/website/models/_data.json index b2898be8a..f7ba16c9f 100644 --- a/website/models/_data.json +++ b/website/models/_data.json @@ -80,6 +80,7 @@ "da": "Danish", "hu": "Hungarian", "pl": "Polish", + "ru": "Russian", "he": "Hebrew", "bn": "Bengali", "id": "Indonesian", diff --git a/website/usage/_models/_languages.jade b/website/usage/_models/_languages.jade index abdad01ad..4337b5b99 100644 --- a/website/usage/_models/_languages.jade +++ b/website/usage/_models/_languages.jade @@ -40,10 +40,13 @@ p +src(gh("spaCy", "spacy/lang/" + code)) #[code lang/#{code}] +infobox("Dependencies") - | Some language tokenizers require external dependencies. To use #[strong Chinese], - | you need to have #[+a("https://github.com/fxsjy/jieba") Jieba] installed. - | The #[strong Japanese] tokenizer requires - | #[+a("https://github.com/mocobeta/janome") Janome]. + .o-block-small Some language tokenizers require external dependencies. + + +list.o-no-block + +item #[strong Chinese]: #[+a("https://github.com/fxsjy/jieba") Jieba] + +item #[strong Japanese]: #[+a("https://github.com/mocobeta/janome") Janome] + +item #[strong Thai]: #[+a("https://github.com/wannaphongcom/pythainlp") pythainlp] + +item #[strong Russian]: #[+a("https://github.com/kmike/pymorphy2") pymorphy2] +h(3, "multi-language") Multi-language support +tag-new(2)