mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-25 17:36:30 +03:00
83 lines
4.5 KiB
Plaintext
83 lines
4.5 KiB
Plaintext
include ../../_includes/_mixins
|
|
|
|
p.u-text-large This tutorial describes how to train new statistical models for spaCy's part-of-speech tagger, named entity recognizer and dependency parser.
|
|
|
|
p I'll start with some quick code examples, that describe how to train each model. I'll then provide a bit of background about the algorithms, and explain how the data and feature templates work.
|
|
|
|
+h(2, "train-pos-tagger") Training the part-of-speech tagger
|
|
|
|
+code('python', 'Simple Example').
|
|
from spacy.vocab import Vocab
|
|
from spacy.pipeline import Tagger
|
|
from spacy.tokens import Doc
|
|
|
|
vocab = Vocab(tag_map={'N': {'pos': 'NOUN'}, 'V': {'pos': 'VERB'}})
|
|
tagger = Tagger(vocab)
|
|
|
|
doc = Doc(vocab, words=['I', 'like', 'stuff'])
|
|
tagger.update(doc, ['N', 'V', 'N'])
|
|
|
|
tagger.model.end_training()
|
|
|
|
p #[+a("https://github.com/" + SOCIAL.github + "/spaCy/examples/training/train_tagger.py") Full example]
|
|
|
|
+h(2, "train-entity") Training the named entity recognizer
|
|
|
|
+code('python', 'Simple Example').
|
|
from spacy.vocab import Vocab
|
|
from spacy.pipeline import EntityRecognizer
|
|
from spacy.tokens import Doc
|
|
from spacy.gold import GoldParse
|
|
|
|
vocab = Vocab()
|
|
entity = EntityRecognizer(vocab, entity_types=['PERSON', 'LOC'])
|
|
|
|
doc = Doc(vocab, words=['Who', 'is', 'Shaka', 'Khan', '?'])
|
|
entity.update(doc, GoldParse(doc, entities=['O', 'O', 'B-PERSON', 'L-PERSON', 'O']))
|
|
|
|
entity.model.end_training()
|
|
|
|
p #[+a("https://github.com/" + SOCIAL.github + "/spaCy/examples/training/train_ner.y") Full example]
|
|
|
|
+h(2, "train-entity") Training the dependency parser
|
|
|
|
+code('python', 'Simple Example').
|
|
from spacy.vocab import Vocab
|
|
from spacy.pipeline import DependencyParser
|
|
from spacy.tokens import Doc
|
|
|
|
vocab = Vocab()
|
|
parser = DependencyParser(vocab, labels=['nsubj', 'compound', 'dobj', 'punct'])
|
|
|
|
doc = Doc(vocab, words=['Who', 'is', 'Shaka', 'Khan', '?'])
|
|
parser.update(doc, GoldParse(doc, heads=[1, 1, 3, 1, 1,], deps=['nsubj', 'ROOT', 'compound', 'dobj', 'punct']))
|
|
|
|
parser.model.end_training()
|
|
|
|
p #[+a("https://github.com/" + SOCIAL.github + "/spaCy/examples/training/train_parser.py") Full example]
|
|
|
|
+h(2, 'feature-templates') Customising the feature extraction
|
|
|
|
p spaCy currently uses linear models for the tagger, parser and entity recognizer, with weights learned using the #[+a("https://explosion.ai/blog/part-of-speech-pos-tagger-in-python") Averaged Perceptron algorithm].
|
|
|
|
p Because it's a linear model, it's important for accuracy to build conjunction features out of the atomic predictors. Let's say you have two atomic predictors asking, "What is the part-of-speech of the previous token?", and "What is the part-of-speech of the previous previous token?". These ppredictors will introduce a number of features, e.g. "Prev-pos=NN", "Prev-pos=VBZ", etc. A conjunction template introduces features such as "Prev-pos=NN&Prev-pos=VBZ".
|
|
|
|
p The feature extraction proceeds in two passes. In the first pass, we fill an array with the values of all of the atomic predictors. In the second pass, we iterate over the feature templates, and fill a small temporary array with the predictors that will be combined into a conjunction feature. Finally, we hash this array into a 64-bit integer, using the MurmurHash algorithm. You can see this at work in the #[+a("https://github.com/" + SOCIAL.github + "/thinc/blob/94dbe06fd3c8f24d86ab0f5c7984e52dbfcdc6cb/thinc/linear/features.pyx") thinc.linear.features] module.
|
|
|
|
p It's very easy to change the feature templates, to create novel combinations of the existing atomic predictors. There's currently no API available to add new atomic predictors, though. You'll have to create a subclass of the model, and write your own #[+code set_featuresC] method.
|
|
|
|
p The feature templates are passed in using the #[+code features] keyword argument to the constructors of the Tagger, DependencyParser and EntityRecognizer:
|
|
|
|
+code('python', 'custom tagger templates').
|
|
from spacy.vocab import Vocab
|
|
from spacy.pipeline import Tagger
|
|
from spacy.tagger import P2_orth, P1_orth
|
|
from spacy.tagger import P2_cluster, P1_cluster, W_orth, N1_orth, N2_orth
|
|
|
|
vocab = Vocab(tag_map={'N': {'pos': 'NOUN'}, 'V': {'pos': 'VERB'}})
|
|
tagger = Tagger(vocab, features=[(P2_orth, P2_cluster), (P1_orth, P1_cluster),
|
|
(P2_orth,), (P1_orth,), (W_orth,),
|
|
(N1_orth,), (N2_orth,)])
|
|
|
|
p Custom feature templates can be passed to the DependencyParser and EntityRecognizer as well, also using the #[+code features] keyword argument of the constructor.
|