From c4f324d5f14c9b216237ebbce6cb691df2e26b65 Mon Sep 17 00:00:00 2001 From: svlandeg Date: Sat, 12 Sep 2020 17:38:54 +0200 Subject: [PATCH 1/5] doc fixes --- website/docs/api/dependencymatcher.md | 2 +- website/docs/api/dependencyparser.md | 2 +- website/docs/api/entitylinker.md | 4 ++-- website/docs/api/entityrecognizer.md | 2 +- website/docs/api/entityruler.md | 2 +- website/docs/api/lemmatizer.md | 2 +- website/docs/api/morphologizer.md | 4 ++-- website/docs/api/phrasematcher.md | 4 ++-- website/docs/api/pipe.md | 4 ++-- website/docs/api/sentencerecognizer.md | 4 ++-- website/docs/api/sentencizer.md | 2 +- website/docs/api/tagger.md | 4 ++-- website/docs/api/textcategorizer.md | 6 +++--- website/docs/api/tok2vec.md | 4 ++-- 14 files changed, 23 insertions(+), 23 deletions(-) diff --git a/website/docs/api/dependencymatcher.md b/website/docs/api/dependencymatcher.md index c90a715d9..356adcda7 100644 --- a/website/docs/api/dependencymatcher.md +++ b/website/docs/api/dependencymatcher.md @@ -183,7 +183,7 @@ will be overwritten. | -------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------- | | `match_id` | An ID for the patterns. ~~str~~ | | `patterns` | A list of match patterns. A pattern consists of a list of dicts, where each dict describes a token in the tree. ~~List[List[Dict[str, Union[str, Dict]]]]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `on_match` | Callback function to act on matches. Takes the arguments `matcher`, `doc`, `i` and `matches`. ~~Optional[Callable[[DependencyMatcher, Doc, int, List[Tuple], Any]]~~ | ## DependencyMatcher.get {#get tag="method"} diff --git a/website/docs/api/dependencyparser.md b/website/docs/api/dependencyparser.md index 674812567..8af4455d3 100644 --- a/website/docs/api/dependencyparser.md +++ b/website/docs/api/dependencyparser.md @@ -217,7 +217,7 @@ model. Delegates to [`predict`](/api/dependencyparser#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | diff --git a/website/docs/api/entitylinker.md b/website/docs/api/entitylinker.md index a9d45d68e..9cb35b487 100644 --- a/website/docs/api/entitylinker.md +++ b/website/docs/api/entitylinker.md @@ -85,7 +85,7 @@ providing custom registered functions. | `vocab` | The shared vocabulary. ~~Vocab~~ | | `model` | The [`Model`](https://thinc.ai/docs/api-model) powering the pipeline component. ~~Model~~ | | `name` | String name of the component instance. Used to add entries to the `losses` during training. ~~str~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `kb_loader` | Function that creates a [`KnowledgeBase`](/api/kb) from a `Vocab` instance. ~~Callable[[Vocab], KnowledgeBase]~~ | | `get_candidates` | Function that generates plausible candidates for a given `Span` object. ~~Callable[[KnowledgeBase, Span], Iterable[Candidate]]~~ | | `labels_discard` | NER labels that will automatically get a `"NIL"` prediction. ~~Iterable[str]~~ | @@ -218,7 +218,7 @@ pipe's entity linking model and context encoder. Delegates to | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | diff --git a/website/docs/api/entityrecognizer.md b/website/docs/api/entityrecognizer.md index 1420aa1a7..8af73f44b 100644 --- a/website/docs/api/entityrecognizer.md +++ b/website/docs/api/entityrecognizer.md @@ -206,7 +206,7 @@ model. Delegates to [`predict`](/api/entityrecognizer#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | diff --git a/website/docs/api/entityruler.md b/website/docs/api/entityruler.md index a6934eeef..7be44bc95 100644 --- a/website/docs/api/entityruler.md +++ b/website/docs/api/entityruler.md @@ -255,7 +255,7 @@ Get all patterns that were added to the entity ruler. | Name | Description | | ----------------- | --------------------------------------------------------------------------------------------------------------------- | -| `matcher` | The underlying matcher used to process token patterns. ~~Matcher~~ | | +| `matcher` | The underlying matcher used to process token patterns. ~~Matcher~~ | | `phrase_matcher` | The underlying phrase matcher, used to process phrase patterns. ~~PhraseMatcher~~ | | `token_patterns` | The token patterns present in the entity ruler, keyed by label. ~~Dict[str, List[Dict[str, Union[str, List[dict]]]]~~ | | `phrase_patterns` | The phrase patterns present in the entity ruler, keyed by label. ~~Dict[str, List[Doc]]~~ | diff --git a/website/docs/api/lemmatizer.md b/website/docs/api/lemmatizer.md index 486410907..f9978dcf9 100644 --- a/website/docs/api/lemmatizer.md +++ b/website/docs/api/lemmatizer.md @@ -81,7 +81,7 @@ shortcut for this and instantiate the component using its string name and | `vocab` | The shared vocabulary. ~~Vocab~~ | | `model` | **Not yet implemented:** The model to use. ~~Model~~ | | `name` | String name of the component instance. Used to add entries to the `losses` during training. ~~str~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | mode | The lemmatizer mode, e.g. `"lookup"` or `"rule"`. Defaults to `"lookup"`. ~~str~~ | | lookups | A lookups object containing the tables such as `"lemma_rules"`, `"lemma_index"`, `"lemma_exc"` and `"lemma_lookup"`. Defaults to `None`. ~~Optional[Lookups]~~ | | overwrite | Whether to overwrite existing lemmas. ~~bool~ | diff --git a/website/docs/api/morphologizer.md b/website/docs/api/morphologizer.md index f2b2f9cc0..e1a166474 100644 --- a/website/docs/api/morphologizer.md +++ b/website/docs/api/morphologizer.md @@ -139,7 +139,7 @@ setting up the label scheme based on the data. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------------------- | | `get_examples` | Function that returns gold-standard annotations in the form of [`Example`](/api/example) objects. ~~Callable[[], Iterable[Example]]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `pipeline` | Optional list of pipeline components that this component is part of. ~~Optional[List[Tuple[str, Callable[[Doc], Doc]]]]~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | **RETURNS** | The optimizer. ~~Optimizer~~ | @@ -196,7 +196,7 @@ Delegates to [`predict`](/api/morphologizer#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | diff --git a/website/docs/api/phrasematcher.md b/website/docs/api/phrasematcher.md index 39e3a298b..47bbdcf6a 100644 --- a/website/docs/api/phrasematcher.md +++ b/website/docs/api/phrasematcher.md @@ -150,9 +150,9 @@ patterns = [nlp("health care reform"), nlp("healthcare reform")] | Name | Description | | -------------- | ---------------------------------------------------------------------------------------------------------------------------------------------------------- | -| `match_id` | str | An ID for the thing you're matching. ~~str~~ | +| `match_id` | An ID for the thing you're matching. ~~str~~ | | | `docs` | `Doc` objects of the phrases to match. ~~List[Doc]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `on_match` | Callback function to act on matches. Takes the arguments `matcher`, `doc`, `i` and `matches`. ~~Optional[Callable[[Matcher, Doc, int, List[tuple], Any]]~~ | ## PhraseMatcher.remove {#remove tag="method" new="2.2"} diff --git a/website/docs/api/pipe.md b/website/docs/api/pipe.md index c8d61a5a9..e4e1e97f1 100644 --- a/website/docs/api/pipe.md +++ b/website/docs/api/pipe.md @@ -187,7 +187,7 @@ predictions and gold-standard annotations, and update the component's model. | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | @@ -211,7 +211,7 @@ the "catastrophic forgetting" problem. This feature is experimental. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------ | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | `losses` | Optional record of the loss during training. Updated using the component name as the key. ~~Optional[Dict[str, float]]~~ | diff --git a/website/docs/api/sentencerecognizer.md b/website/docs/api/sentencerecognizer.md index ca19327bb..acf94fb8e 100644 --- a/website/docs/api/sentencerecognizer.md +++ b/website/docs/api/sentencerecognizer.md @@ -192,7 +192,7 @@ Delegates to [`predict`](/api/sentencerecognizer#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | @@ -216,7 +216,7 @@ the "catastrophic forgetting" problem. This feature is experimental. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------ | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | `losses` | Optional record of the loss during training. Updated using the component name as the key. ~~Optional[Dict[str, float]]~~ | diff --git a/website/docs/api/sentencizer.md b/website/docs/api/sentencizer.md index c435acdcb..ae31e4ddf 100644 --- a/website/docs/api/sentencizer.md +++ b/website/docs/api/sentencizer.md @@ -53,7 +53,7 @@ Initialize the sentencizer. | Name | Description | | -------------- | ----------------------------------------------------------------------------------------------------------------------- | -| _keyword-only_ | | | +| _keyword-only_ | | | `punct_chars` | Optional custom list of punctuation characters that mark sentence ends. See below for defaults. ~~Optional[List[str]]~~ | ```python diff --git a/website/docs/api/tagger.md b/website/docs/api/tagger.md index d83a77357..d428d376e 100644 --- a/website/docs/api/tagger.md +++ b/website/docs/api/tagger.md @@ -190,7 +190,7 @@ Delegates to [`predict`](/api/tagger#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | @@ -214,7 +214,7 @@ the "catastrophic forgetting" problem. This feature is experimental. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------ | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | `losses` | Optional record of the loss during training. Updated using the component name as the key. ~~Optional[Dict[str, float]]~~ | diff --git a/website/docs/api/textcategorizer.md b/website/docs/api/textcategorizer.md index cc20d6fd2..75bd4a5bf 100644 --- a/website/docs/api/textcategorizer.md +++ b/website/docs/api/textcategorizer.md @@ -198,7 +198,7 @@ Delegates to [`predict`](/api/textcategorizer#predict) and | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | @@ -222,7 +222,7 @@ the "catastrophic forgetting" problem. This feature is experimental. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------ | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | `losses` | Optional record of the loss during training. Updated using the component name as the key. ~~Optional[Dict[str, float]]~~ | @@ -260,7 +260,7 @@ Score a batch of examples. | Name | Description | | ---------------- | -------------------------------------------------------------------------------------------------------------------- | | `examples` | The examples to score. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `positive_label` | Optional positive label. ~~Optional[str]~~ | | **RETURNS** | The scores, produced by [`Scorer.score_cats`](/api/scorer#score_cats). ~~Dict[str, Union[float, Dict[str, float]]]~~ | diff --git a/website/docs/api/tok2vec.md b/website/docs/api/tok2vec.md index 6f13a17a5..5c7214edc 100644 --- a/website/docs/api/tok2vec.md +++ b/website/docs/api/tok2vec.md @@ -144,7 +144,7 @@ setting up the label scheme based on the data. | Name | Description | | -------------- | ------------------------------------------------------------------------------------------------------------------------------------- | | `get_examples` | Function that returns gold-standard annotations in the form of [`Example`](/api/example) objects. ~~Callable[[], Iterable[Example]]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `pipeline` | Optional list of pipeline components that this component is part of. ~~Optional[List[Tuple[str, Callable[[Doc], Doc]]]]~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | | **RETURNS** | The optimizer. ~~Optimizer~~ | @@ -200,7 +200,7 @@ Delegates to [`predict`](/api/tok2vec#predict). | Name | Description | | ----------------- | ---------------------------------------------------------------------------------------------------------------------------------- | | `examples` | A batch of [`Example`](/api/example) objects to learn from. ~~Iterable[Example]~~ | -| _keyword-only_ | | | +| _keyword-only_ | | | `drop` | The dropout rate. ~~float~~ | | `set_annotations` | Whether or not to update the `Example` objects with the predictions, delegating to [`set_annotations`](#set_annotations). ~~bool~~ | | `sgd` | An optimizer. Will be created via [`create_optimizer`](#create_optimizer) if not set. ~~Optional[Optimizer]~~ | From 80754d7065826e7e4275fafc12b8f979a43d7fd3 Mon Sep 17 00:00:00 2001 From: Ines Montani Date: Mon, 14 Sep 2020 10:29:06 +0200 Subject: [PATCH 2/5] Update README.md [ci skip] --- README.md | 78 ++++++++++++++++++++++++------------------------------- 1 file changed, 34 insertions(+), 44 deletions(-) diff --git a/README.md b/README.md index cef2a1fdd..d23051af0 100644 --- a/README.md +++ b/README.md @@ -4,17 +4,19 @@ spaCy is a library for advanced Natural Language Processing in Python and Cython. It's built on the very latest research, and was designed from day one to -be used in real products. spaCy comes with -[pretrained statistical models](https://spacy.io/models) and word vectors, and -currently supports tokenization for **60+ languages**. It features +be used in real products. + +spaCy comes with +[pretrained pipelines](https://spacy.io/models) and vectors, and +currently supports tokenization for **59+ languages**. It features state-of-the-art speed, convolutional **neural network models** for tagging, -parsing and **named entity recognition** and easy **deep learning** integration. -It's commercial open-source software, released under the MIT license. +parsing, **named entity recognition**, **text classification** and more, multi-task learning with pretrained **transformers** like BERT, as well as a production-ready training system and easy model packaging, deployment and workflow management. +spaCy is commercial open-source software, released under the MIT license. ๐Ÿ’ซ **Version 2.3 out now!** [Check out the release notes here.](https://github.com/explosion/spaCy/releases) -[![Azure Pipelines]()](https://dev.azure.com/explosion-ai/public/_build?definitionId=8) +[![Azure Pipelines](https://img.shields.io/azure-devops/build/explosion-ai/public/8/master.svg?logo=azure-pipelines&style=flat-square&label=build)](https://dev.azure.com/explosion-ai/public/_build?definitionId=8) [![Current Release Version](https://img.shields.io/github/release/explosion/spacy.svg?style=flat-square&logo=github)](https://github.com/explosion/spaCy/releases) [![pypi Version](https://img.shields.io/pypi/v/spacy.svg?style=flat-square&logo=pypi&logoColor=white)](https://pypi.org/project/spacy/) [![conda Version](https://img.shields.io/conda/vn/conda-forge/spacy.svg?style=flat-square&logo=conda-forge&logoColor=white)](https://anaconda.org/conda-forge/spacy) @@ -31,7 +33,7 @@ It's commercial open-source software, released under the MIT license. | --------------- | -------------------------------------------------------------- | | [spaCy 101] | New to spaCy? Here's everything you need to know! | | [Usage Guides] | How to use spaCy and its features. | -| [New in v2.3] | New features, backwards incompatibilities and migration guide. | +| [New in v3.0] | New features, backwards incompatibilities and migration guide. | | [API Reference] | The detailed reference for spaCy's API. | | [Models] | Download statistical language models for spaCy. | | [Universe] | Libraries, extensions, demos, books and courses. | @@ -39,7 +41,7 @@ It's commercial open-source software, released under the MIT license. | [Contribute] | How to contribute to the spaCy project and code base. | [spacy 101]: https://spacy.io/usage/spacy-101 -[new in v2.3]: https://spacy.io/usage/v2-3 +[new in v3.0]: https://spacy.io/usage/v3 [usage guides]: https://spacy.io/usage/ [api reference]: https://spacy.io/api/ [models]: https://spacy.io/models @@ -56,34 +58,29 @@ be able to provide individual support via email. We also believe that help is much more valuable if it's shared publicly, so that more people can benefit from it. -| Type | Platforms | -| ------------------------ | ------------------------------------------------------ | -| ๐Ÿšจ **Bug Reports** | [GitHub Issue Tracker] | -| ๐ŸŽ **Feature Requests** | [GitHub Issue Tracker] | -| ๐Ÿ‘ฉโ€๐Ÿ’ป **Usage Questions** | [Stack Overflow] ยท [Gitter Chat] ยท [Reddit User Group] | -| ๐Ÿ—ฏ **General Discussion** | [Gitter Chat] ยท [Reddit User Group] | +| Type | Platforms | +| ----------------------- | ---------------------- | +| ๐Ÿšจ **Bug Reports** | [GitHub Issue Tracker] | +| ๐ŸŽ **Feature Requests** | [GitHub Issue Tracker] | +| ๐Ÿ‘ฉโ€๐Ÿ’ป **Usage Questions** | [Stack Overflow] | [github issue tracker]: https://github.com/explosion/spaCy/issues [stack overflow]: https://stackoverflow.com/questions/tagged/spacy -[gitter chat]: https://gitter.im/explosion/spaCy -[reddit user group]: https://www.reddit.com/r/spacynlp ## Features -- Non-destructive **tokenization** -- **Named entity** recognition -- Support for **50+ languages** -- pretrained [statistical models](https://spacy.io/models) and word vectors +- Support for **59+ languages** +- **Trained pipelines** +- Multi-task learning with pretrained **transformers** like BERT +- Pretrained **word vectors** - State-of-the-art speed -- Easy **deep learning** integration -- Part-of-speech tagging -- Labelled dependency parsing -- Syntax-driven sentence segmentation +- Production-ready **training system** +- Linguistically-motivated **tokenization** +- Components for named **entity recognition**, part-of-speech-tagging, dependency parsing, sentence segmentation, **text classification**, lemmatization, morphological analysis, entity linking and more +- Easily extensible with **custom components** and attributes +- Support for custom models in **PyTorch**, **TensorFlow** and other frameworks - Built in **visualizers** for syntax and NER -- Convenient string-to-hash mapping -- Export to numpy data arrays -- Efficient binary serialization -- Easy **model packaging** and deployment +- Easy **model packaging**, deployment and workflow management - Robust, rigorously evaluated accuracy ๐Ÿ“– **For more details, see the @@ -102,13 +99,6 @@ For detailed installation instructions, see the [pip]: https://pypi.org/project/spacy/ [conda]: https://anaconda.org/conda-forge/spacy -> โš ๏ธ **Important note for Python 3.8:** We can't yet ship pre-compiled binary -> wheels for spaCy that work on Python 3.8, as we're still waiting for our CI -> providers and other tooling to support it. This means that in order to run -> spaCy on Python 3.8, you'll need [a compiler installed](#source) and compile -> the library and its Cython dependencies locally. If this is causing problems -> for you, the easiest solution is to **use Python 3.7** in the meantime. - ### pip Using pip, spaCy releases are available as source packages and binary wheels (as @@ -164,26 +154,26 @@ If you've trained your own models, keep in mind that your training and runtime inputs must match. After updating spaCy, we recommend **retraining your models** with the new version. -๐Ÿ“– **For details on upgrading from spaCy 1.x to spaCy 2.x, see the -[migration guide](https://spacy.io/usage/v2#migrating).** +๐Ÿ“– **For details on upgrading from spaCy 2.x to spaCy 3.x, see the +[migration guide](https://spacy.io/usage/v3#migrating).** ## Download models -As of v1.7.0, models for spaCy can be installed as **Python packages**. This +Trained pipelines for spaCy can be installed as **Python packages**. This means that they're a component of your application, just like any other module. Models can be installed using spaCy's `download` command, or manually by pointing pip to a path or URL. -| Documentation | | -| ---------------------- | ------------------------------------------------------------- | -| [Available Models] | Detailed model descriptions, accuracy figures and benchmarks. | -| [Models Documentation] | Detailed usage instructions. | +| Documentation | | +| ---------------------- | ---------------------------------------------------------------- | +| [Available Pipelines] | Detailed pipeline descriptions, accuracy figures and benchmarks. | +| [Models Documentation] | Detailed usage instructions. | -[available models]: https://spacy.io/models +[available pipelines]: https://spacy.io/models [models documentation]: https://spacy.io/docs/usage/models ```bash -# download best-matching version of specific model for your spaCy installation +# Download best-matching version of specific model for your spaCy installation python -m spacy download en_core_web_sm # pip install .tar.gz archive from path or URL From 35156429c44c2bc6b2894470c0bf1a6dcfab1866 Mon Sep 17 00:00:00 2001 From: Ines Montani Date: Mon, 14 Sep 2020 10:34:50 +0200 Subject: [PATCH 3/5] Update docs [ci skip] --- website/src/components/quickstart.js | 5 +++-- website/src/widgets/quickstart-training.js | 2 ++ 2 files changed, 5 insertions(+), 2 deletions(-) diff --git a/website/src/components/quickstart.js b/website/src/components/quickstart.js index 6a335d4a0..64f828c2f 100644 --- a/website/src/components/quickstart.js +++ b/website/src/components/quickstart.js @@ -27,6 +27,7 @@ const Quickstart = ({ hidePrompts, small, codeLang, + Container = Section, children, }) => { const contentRef = useRef() @@ -83,7 +84,7 @@ const Quickstart = ({ }, [data, initialized]) return !data.length ? null : ( -
+
{title && (

@@ -249,7 +250,7 @@ const Quickstart = ({ {showCopy &&