mirror of
https://github.com/explosion/spaCy.git
synced 2025-06-05 21:53:05 +03:00
Merge pull request #1367 from h4iku/fix-spacy101-typo
Fix small typos in spaCy 101 usage document
This commit is contained in:
commit
55b758154d
|
@ -22,7 +22,7 @@ p
|
||||||
|
|
||||||
+aside-code.
|
+aside-code.
|
||||||
from spacy.lang.en import English
|
from spacy.lang.en import English
|
||||||
from spacy.lang.en import German
|
from spacy.lang.de import German
|
||||||
|
|
||||||
nlp_en = English() # includes English data
|
nlp_en = English() # includes English data
|
||||||
nlp_de = German() # includes German data
|
nlp_de = German() # includes German data
|
||||||
|
|
|
@ -112,4 +112,4 @@ p
|
||||||
| only works if you actually #[em know] that the document contains that
|
| only works if you actually #[em know] that the document contains that
|
||||||
| word. To prevent this problem, spaCy will also export the #[code Vocab]
|
| word. To prevent this problem, spaCy will also export the #[code Vocab]
|
||||||
| when you save a #[code Doc] or #[code nlp] object. This will give you
|
| when you save a #[code Doc] or #[code nlp] object. This will give you
|
||||||
| the object and its encoded annotations, plus they "key" to decode it.
|
| the object and its encoded annotations, plus the "key" to decode it.
|
||||||
|
|
|
@ -65,7 +65,7 @@ p
|
||||||
| not designed specifically for chat bots, and only provides the
|
| not designed specifically for chat bots, and only provides the
|
||||||
| underlying text processing capabilities.
|
| underlying text processing capabilities.
|
||||||
+item #[strong spaCy is not research software].
|
+item #[strong spaCy is not research software].
|
||||||
| It's is built on the latest research, but it's designed to get
|
| It's built on the latest research, but it's designed to get
|
||||||
| things done. This leads to fairly different design decisions than
|
| things done. This leads to fairly different design decisions than
|
||||||
| #[+a("https://github./nltk/nltk") NLTK]
|
| #[+a("https://github./nltk/nltk") NLTK]
|
||||||
| or #[+a("https://stanfordnlp.github.io/CoreNLP/") CoreNLP], which were
|
| or #[+a("https://stanfordnlp.github.io/CoreNLP/") CoreNLP], which were
|
||||||
|
@ -87,7 +87,7 @@ p
|
||||||
|
|
||||||
+aside
|
+aside
|
||||||
| If one of spaCy's functionalities #[strong needs a model], it means that
|
| If one of spaCy's functionalities #[strong needs a model], it means that
|
||||||
| you need to have one our the available
|
| you need to have one of the available
|
||||||
| #[+a("/docs/usage/models") statistical models] installed. Models are used
|
| #[+a("/docs/usage/models") statistical models] installed. Models are used
|
||||||
| to #[strong predict] linguistic annotations – for example, if a word is
|
| to #[strong predict] linguistic annotations – for example, if a word is
|
||||||
| a verb or a noun.
|
| a verb or a noun.
|
||||||
|
|
Loading…
Reference in New Issue
Block a user