2bcceb80c4
* Refactor the Scorer to improve flexibility Refactor the `Scorer` to improve flexibility for arbitrary pipeline components. * Individual pipeline components provide their own `evaluate` methods that score a list of `Example`s and return a dictionary of scores * `Scorer` is initialized either: * with a provided pipeline containing components to be scored * with a default pipeline containing the built-in statistical components (senter, tagger, morphologizer, parser, ner) * `Scorer.score` evaluates a list of `Example`s and returns a dictionary of scores referring to the scores provided by the components in the pipeline Significant differences: * `tags_acc` is renamed to `tag_acc` to be consistent with `token_acc` and the new `morph_acc`, `pos_acc`, and `lemma_acc` * Scoring is no longer cumulative: `Scorer.score` scores a list of examples rather than a single example and does not retain any state about previously scored examples * PRF values in the returned scores are no longer multiplied by 100 * Add kwargs to Morphologizer.evaluate * Create generalized scoring methods in Scorer * Generalized static scoring methods are added to `Scorer` * Methods require an attribute (either on Token or Doc) that is used to key the returned scores Naming differences: * `uas`, `las`, and `las_per_type` in the scores dict are renamed to `dep_uas`, `dep_las`, and `dep_las_per_type` Scoring differences: * `Doc.sents` is now scored as spans rather than on sentence-initial token positions so that `Doc.sents` and `Doc.ents` can be scored with the same method (this lowers scores since a single incorrect sentence start results in two incorrect spans) * Simplify / extend hasattr check for eval method * Add hasattr check to tokenizer scoring * Simplify to hasattr check for component scoring * Reset Example alignment if docs are set Reset the Example alignment if either doc is set in case the tokenization has changed. * Add PRF tokenization scoring for tokens as spans Add PRF scores for tokens as character spans. The scores are: * token_acc: # correct tokens / # gold tokens * token_p/r/f: PRF for (token.idx, token.idx + len(token)) * Add docstring to Scorer.score_tokenization * Rename component.evaluate() to component.score() * Update Scorer API docs * Update scoring for positive_label in textcat * Fix TextCategorizer.score kwargs * Update Language.evaluate docs * Update score names in default config |
||
---|---|---|
.buildkite | ||
.github | ||
bin | ||
examples | ||
include | ||
spacy | ||
website | ||
.gitignore | ||
azure-pipelines.yml | ||
CITATION | ||
CONTRIBUTING.md | ||
fabfile.py | ||
LICENSE | ||
Makefile | ||
MANIFEST.in | ||
netlify.toml | ||
pyproject.toml | ||
README.md | ||
requirements.txt | ||
setup.cfg | ||
setup.py |
spaCy: Industrial-strength NLP
spaCy is a library for advanced Natural Language Processing in Python and Cython. It's built on the very latest research, and was designed from day one to be used in real products. spaCy comes with pretrained statistical models and word vectors, and currently supports tokenization for 60+ languages. It features state-of-the-art speed, convolutional neural network models for tagging, parsing and named entity recognition and easy deep learning integration. It's commercial open-source software, released under the MIT license.
💫 Version 2.3 out now! Check out the release notes here.
📖 Documentation
Documentation | |
---|---|
spaCy 101 | New to spaCy? Here's everything you need to know! |
Usage Guides | How to use spaCy and its features. |
New in v2.3 | New features, backwards incompatibilities and migration guide. |
API Reference | The detailed reference for spaCy's API. |
Models | Download statistical language models for spaCy. |
Universe | Libraries, extensions, demos, books and courses. |
Changelog | Changes and version history. |
Contribute | How to contribute to the spaCy project and code base. |
💬 Where to ask questions
The spaCy project is maintained by @honnibal and @ines, along with core contributors @svlandeg and @adrianeboyd. Please understand that we won't be able to provide individual support via email. We also believe that help is much more valuable if it's shared publicly, so that more people can benefit from it.
Type | Platforms |
---|---|
🚨 Bug Reports | GitHub Issue Tracker |
🎁 Feature Requests | GitHub Issue Tracker |
👩💻 Usage Questions | Stack Overflow · Gitter Chat · Reddit User Group |
🗯 General Discussion | Gitter Chat · Reddit User Group |
Features
- Non-destructive tokenization
- Named entity recognition
- Support for 50+ languages
- pretrained statistical models and word vectors
- State-of-the-art speed
- Easy deep learning integration
- Part-of-speech tagging
- Labelled dependency parsing
- Syntax-driven sentence segmentation
- Built in visualizers for syntax and NER
- Convenient string-to-hash mapping
- Export to numpy data arrays
- Efficient binary serialization
- Easy model packaging and deployment
- Robust, rigorously evaluated accuracy
📖 For more details, see the facts, figures and benchmarks.
Install spaCy
For detailed installation instructions, see the documentation.
- Operating system: macOS / OS X · Linux · Windows (Cygwin, MinGW, Visual Studio)
- Python version: Python 3.6+ (only 64 bit)
- Package managers: pip · conda (via
conda-forge
)
⚠️ Important note for Python 3.8: We can't yet ship pre-compiled binary wheels for spaCy that work on Python 3.8, as we're still waiting for our CI providers and other tooling to support it. This means that in order to run spaCy on Python 3.8, you'll need a compiler installed and compile the library and its Cython dependencies locally. If this is causing problems for you, the easiest solution is to use Python 3.7 in the meantime.
pip
Using pip, spaCy releases are available as source packages and binary wheels (as
of v2.0.13
).
pip install spacy
To install additional data tables for lemmatization and normalization in
spaCy v2.2+ you can run pip install spacy[lookups]
or install
spacy-lookups-data
separately. The lookups package is needed to create blank models with
lemmatization data for v2.2+ plus normalization data for v2.3+, and to
lemmatize in languages that don't yet come with pretrained models and aren't
powered by third-party libraries.
When using pip it is generally recommended to install packages in a virtual environment to avoid modifying system state:
python -m venv .env
source .env/bin/activate
pip install spacy
conda
Thanks to our great community, we've finally re-added conda support. You can now
install spaCy via conda-forge
:
conda install -c conda-forge spacy
For the feedstock including the build recipe and configuration, check out this repository. Improvements and pull requests to the recipe and setup are always appreciated.
Updating spaCy
Some updates to spaCy may require downloading new statistical models. If you're
running spaCy v2.0 or higher, you can use the validate
command to check if
your installed models are compatible and if not, print details on how to update
them:
pip install -U spacy
python -m spacy validate
If you've trained your own models, keep in mind that your training and runtime inputs must match. After updating spaCy, we recommend retraining your models with the new version.
📖 For details on upgrading from spaCy 1.x to spaCy 2.x, see the migration guide.
Download models
As of v1.7.0, models for spaCy can be installed as Python packages. This
means that they're a component of your application, just like any other module.
Models can be installed using spaCy's download
command, or manually by
pointing pip to a path or URL.
Documentation | |
---|---|
Available Models | Detailed model descriptions, accuracy figures and benchmarks. |
Models Documentation | Detailed usage instructions. |
# download best-matching version of specific model for your spaCy installation
python -m spacy download en_core_web_sm
# pip install .tar.gz archive from path or URL
pip install /Users/you/en_core_web_sm-2.2.0.tar.gz
pip install https://github.com/explosion/spacy-models/releases/download/en_core_web_sm-2.2.0/en_core_web_sm-2.2.0.tar.gz
Loading and using models
To load a model, use spacy.load()
with the model name or a
path to the model data directory.
import spacy
nlp = spacy.load("en_core_web_sm")
doc = nlp("This is a sentence.")
You can also import
a model directly via its full name and then call its
load()
method with no arguments.
import spacy
import en_core_web_sm
nlp = en_core_web_sm.load()
doc = nlp("This is a sentence.")
📖 For more info and examples, check out the models documentation.
Compile from source
The other way to install spaCy is to clone its GitHub repository and build it from source. That is the common way if you want to make changes to the code base. You'll need to make sure that you have a development environment consisting of a Python distribution including header files, a compiler, pip, virtualenv and git installed. The compiler part is the trickiest. How to do that depends on your system. See notes on Ubuntu, OS X and Windows for details.
# make sure you are using the latest pip
python -m pip install -U pip
git clone https://github.com/explosion/spaCy
cd spaCy
python -m venv .env
source .env/bin/activate
export PYTHONPATH=`pwd`
pip install -r requirements.txt
python setup.py build_ext --inplace
Compared to regular install via pip, requirements.txt additionally installs developer dependencies such as Cython. For more details and instructions, see the documentation on compiling spaCy from source and the quickstart widget to get the right commands for your platform and Python version.
Ubuntu
Install system-level dependencies via apt-get
:
sudo apt-get install build-essential python-dev git
macOS / OS X
Install a recent version of XCode, including the so-called "Command Line Tools". macOS and OS X ship with Python and git preinstalled.
Windows
Install a version of the Visual C++ Build Tools or Visual Studio Express that matches the version that was used to compile your Python interpreter.
Run tests
spaCy comes with an extensive test suite. In order to run the
tests, you'll usually want to clone the repository and build spaCy from source.
This will also install the required development dependencies and test utilities
defined in the requirements.txt
.
Alternatively, you can find out where spaCy is installed and run pytest
on
that directory. Don't forget to also install the test utilities via spaCy's
requirements.txt
:
python -c "import os; import spacy; print(os.path.dirname(spacy.__file__))"
pip install -r path/to/requirements.txt
python -m pytest <spacy-directory>
See the documentation for more details and examples.