💫 Update website (#3285)
<!--- Provide a general summary of your changes in the title. -->
## Description
The new website is implemented using [Gatsby](https://www.gatsbyjs.org) with [Remark](https://github.com/remarkjs/remark) and [MDX](https://mdxjs.com/). This allows authoring content in **straightforward Markdown** without the usual limitations. Standard elements can be overwritten with powerful [React](http://reactjs.org/) components and wherever Markdown syntax isn't enough, JSX components can be used. Hopefully, this update will also make it much easier to contribute to the docs. Once this PR is merged, I'll implement auto-deployment via [Netlify](https://netlify.com) on a specific branch (to avoid building the website on every PR). There's a bunch of other cool stuff that the new setup will allow us to do – including writing front-end tests, service workers, offline support, implementing a search and so on.
This PR also includes various new docs pages and content.
Resolves #3270. Resolves #3222. Resolves #2947. Resolves #2837.
### Types of change
enhancement
## Checklist
<!--- Before you submit the PR, go over this checklist and make sure you can
tick off all the boxes. [] -> [x] -->
- [x] I have submitted the spaCy Contributor Agreement.
- [x] I ran the tests, and all new and existing tests passed.
- [x] My changes don't require a change to the documentation, or if they do, I've added all required information.
2019-02-17 21:31:19 +03:00
|
|
|
|
After tokenization, spaCy can **parse** and **tag** a given `Doc`. This is where
|
|
|
|
|
the statistical model comes in, which enables spaCy to **make a prediction** of
|
|
|
|
|
which tag or label most likely applies in this context. A model consists of
|
|
|
|
|
binary data and is produced by showing a system enough examples for it to make
|
|
|
|
|
predictions that generalize across the language – for example, a word following
|
|
|
|
|
"the" in English is most likely a noun.
|
|
|
|
|
|
|
|
|
|
Linguistic annotations are available as
|
|
|
|
|
[`Token` attributes](/api/token#attributes). Like many NLP libraries, spaCy
|
|
|
|
|
**encodes all strings to hash values** to reduce memory usage and improve
|
|
|
|
|
efficiency. So to get the readable string representation of an attribute, we
|
|
|
|
|
need to add an underscore `_` to its name:
|
|
|
|
|
|
|
|
|
|
```python
|
|
|
|
|
### {executable="true"}
|
|
|
|
|
import spacy
|
|
|
|
|
|
2019-09-12 17:11:15 +03:00
|
|
|
|
nlp = spacy.load("en_core_web_sm")
|
|
|
|
|
doc = nlp("Apple is looking at buying U.K. startup for $1 billion")
|
💫 Update website (#3285)
<!--- Provide a general summary of your changes in the title. -->
## Description
The new website is implemented using [Gatsby](https://www.gatsbyjs.org) with [Remark](https://github.com/remarkjs/remark) and [MDX](https://mdxjs.com/). This allows authoring content in **straightforward Markdown** without the usual limitations. Standard elements can be overwritten with powerful [React](http://reactjs.org/) components and wherever Markdown syntax isn't enough, JSX components can be used. Hopefully, this update will also make it much easier to contribute to the docs. Once this PR is merged, I'll implement auto-deployment via [Netlify](https://netlify.com) on a specific branch (to avoid building the website on every PR). There's a bunch of other cool stuff that the new setup will allow us to do – including writing front-end tests, service workers, offline support, implementing a search and so on.
This PR also includes various new docs pages and content.
Resolves #3270. Resolves #3222. Resolves #2947. Resolves #2837.
### Types of change
enhancement
## Checklist
<!--- Before you submit the PR, go over this checklist and make sure you can
tick off all the boxes. [] -> [x] -->
- [x] I have submitted the spaCy Contributor Agreement.
- [x] I ran the tests, and all new and existing tests passed.
- [x] My changes don't require a change to the documentation, or if they do, I've added all required information.
2019-02-17 21:31:19 +03:00
|
|
|
|
|
|
|
|
|
for token in doc:
|
|
|
|
|
print(token.text, token.lemma_, token.pos_, token.tag_, token.dep_,
|
|
|
|
|
token.shape_, token.is_alpha, token.is_stop)
|
|
|
|
|
```
|
|
|
|
|
|
|
|
|
|
> - **Text:** The original word text.
|
|
|
|
|
> - **Lemma:** The base form of the word.
|
|
|
|
|
> - **POS:** The simple part-of-speech tag.
|
|
|
|
|
> - **Tag:** The detailed part-of-speech tag.
|
|
|
|
|
> - **Dep:** Syntactic dependency, i.e. the relation between tokens.
|
|
|
|
|
> - **Shape:** The word shape – capitalization, punctuation, digits.
|
|
|
|
|
> - **is alpha:** Is the token an alpha character?
|
|
|
|
|
> - **is stop:** Is the token part of a stop list, i.e. the most common words of
|
|
|
|
|
> the language?
|
|
|
|
|
|
|
|
|
|
| Text | Lemma | POS | Tag | Dep | Shape | alpha | stop |
|
|
|
|
|
| ------- | ------- | ------- | ----- | ---------- | ------- | ------- | ------- |
|
|
|
|
|
| Apple | apple | `PROPN` | `NNP` | `nsubj` | `Xxxxx` | `True` | `False` |
|
|
|
|
|
| is | be | `VERB` | `VBZ` | `aux` | `xx` | `True` | `True` |
|
|
|
|
|
| looking | look | `VERB` | `VBG` | `ROOT` | `xxxx` | `True` | `False` |
|
|
|
|
|
| at | at | `ADP` | `IN` | `prep` | `xx` | `True` | `True` |
|
|
|
|
|
| buying | buy | `VERB` | `VBG` | `pcomp` | `xxxx` | `True` | `False` |
|
|
|
|
|
| U.K. | u.k. | `PROPN` | `NNP` | `compound` | `X.X.` | `False` | `False` |
|
|
|
|
|
| startup | startup | `NOUN` | `NN` | `dobj` | `xxxx` | `True` | `False` |
|
|
|
|
|
| for | for | `ADP` | `IN` | `prep` | `xxx` | `True` | `True` |
|
|
|
|
|
| \$ | \$ | `SYM` | `$` | `quantmod` | `$` | `False` | `False` |
|
|
|
|
|
| 1 | 1 | `NUM` | `CD` | `compound` | `d` | `False` | `False` |
|
|
|
|
|
| billion | billion | `NUM` | `CD` | `probj` | `xxxx` | `True` | `False` |
|
|
|
|
|
|
|
|
|
|
> #### Tip: Understanding tags and labels
|
|
|
|
|
>
|
|
|
|
|
> Most of the tags and labels look pretty abstract, and they vary between
|
|
|
|
|
> languages. `spacy.explain` will show you a short description – for example,
|
|
|
|
|
> `spacy.explain("VBZ")` returns "verb, 3rd person singular present".
|
|
|
|
|
|
|
|
|
|
Using spaCy's built-in [displaCy visualizer](/usage/visualizers), here's what
|
|
|
|
|
our example sentence and its dependencies look like:
|
|
|
|
|
|
|
|
|
|
import DisplaCyLongHtml from 'images/displacy-long.html'; import { Iframe } from
|
|
|
|
|
'components/embed'
|
|
|
|
|
|
|
|
|
|
<Iframe title="displaCy visualization of dependencies and entities" html={DisplaCyLongHtml} height={450} />
|