mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-25 17:36:30 +03:00
Merge remote-tracking branch 'upstream/master' into feature/nel-fixes
This commit is contained in:
commit
ad65171837
106
.github/contributors/FallakAsad.md
vendored
Normal file
106
.github/contributors/FallakAsad.md
vendored
Normal file
|
@ -0,0 +1,106 @@
|
||||||
|
# spaCy contributor agreement
|
||||||
|
|
||||||
|
This spaCy Contributor Agreement (**"SCA"**) is based on the
|
||||||
|
[Oracle Contributor Agreement](http://www.oracle.com/technetwork/oca-405177.pdf).
|
||||||
|
The SCA applies to any contribution that you make to any product or project
|
||||||
|
managed by us (the **"project"**), and sets out the intellectual property rights
|
||||||
|
you grant to us in the contributed materials. The term **"us"** shall mean
|
||||||
|
[ExplosionAI GmbH](https://explosion.ai/legal). The term
|
||||||
|
**"you"** shall mean the person or entity identified below.
|
||||||
|
|
||||||
|
If you agree to be bound by these terms, fill in the information requested
|
||||||
|
below and include the filled-in version with your first pull request, under the
|
||||||
|
folder [`.github/contributors/`](/.github/contributors/). The name of the file
|
||||||
|
should be your GitHub username, with the extension `.md`. For example, the user
|
||||||
|
example_user would create the file `.github/contributors/example_user.md`.
|
||||||
|
|
||||||
|
Read this agreement carefully before signing. These terms and conditions
|
||||||
|
constitute a binding legal agreement.
|
||||||
|
|
||||||
|
## Contributor Agreement
|
||||||
|
|
||||||
|
1. The term "contribution" or "contributed materials" means any source code,
|
||||||
|
object code, patch, tool, sample, graphic, specification, manual,
|
||||||
|
documentation, or any other material posted or submitted by you to the project.
|
||||||
|
|
||||||
|
2. With respect to any worldwide copyrights, or copyright applications and
|
||||||
|
registrations, in your contribution:
|
||||||
|
|
||||||
|
* you hereby assign to us joint ownership, and to the extent that such
|
||||||
|
assignment is or becomes invalid, ineffective or unenforceable, you hereby
|
||||||
|
grant to us a perpetual, irrevocable, non-exclusive, worldwide, no-charge,
|
||||||
|
royalty-free, unrestricted license to exercise all rights under those
|
||||||
|
copyrights. This includes, at our option, the right to sublicense these same
|
||||||
|
rights to third parties through multiple levels of sublicensees or other
|
||||||
|
licensing arrangements;
|
||||||
|
|
||||||
|
* you agree that each of us can do all things in relation to your
|
||||||
|
contribution as if each of us were the sole owners, and if one of us makes
|
||||||
|
a derivative work of your contribution, the one who makes the derivative
|
||||||
|
work (or has it made will be the sole owner of that derivative work;
|
||||||
|
|
||||||
|
* you agree that you will not assert any moral rights in your contribution
|
||||||
|
against us, our licensees or transferees;
|
||||||
|
|
||||||
|
* you agree that we may register a copyright in your contribution and
|
||||||
|
exercise all ownership rights associated with it; and
|
||||||
|
|
||||||
|
* you agree that neither of us has any duty to consult with, obtain the
|
||||||
|
consent of, pay or render an accounting to the other for any use or
|
||||||
|
distribution of your contribution.
|
||||||
|
|
||||||
|
3. With respect to any patents you own, or that you can license without payment
|
||||||
|
to any third party, you hereby grant to us a perpetual, irrevocable,
|
||||||
|
non-exclusive, worldwide, no-charge, royalty-free license to:
|
||||||
|
|
||||||
|
* make, have made, use, sell, offer to sell, import, and otherwise transfer
|
||||||
|
your contribution in whole or in part, alone or in combination with or
|
||||||
|
included in any product, work or materials arising out of the project to
|
||||||
|
which your contribution was submitted, and
|
||||||
|
|
||||||
|
* at our option, to sublicense these same rights to third parties through
|
||||||
|
multiple levels of sublicensees or other licensing arrangements.
|
||||||
|
|
||||||
|
4. Except as set out above, you keep all right, title, and interest in your
|
||||||
|
contribution. The rights that you grant to us under these terms are effective
|
||||||
|
on the date you first submitted a contribution to us, even if your submission
|
||||||
|
took place before the date you sign these terms.
|
||||||
|
|
||||||
|
5. You covenant, represent, warrant and agree that:
|
||||||
|
|
||||||
|
* Each contribution that you submit is and shall be an original work of
|
||||||
|
authorship and you can legally grant the rights set out in this SCA;
|
||||||
|
|
||||||
|
* to the best of your knowledge, each contribution will not violate any
|
||||||
|
third party's copyrights, trademarks, patents, or other intellectual
|
||||||
|
property rights; and
|
||||||
|
|
||||||
|
* each contribution shall be in compliance with U.S. export control laws and
|
||||||
|
other applicable export and import laws. You agree to notify us if you
|
||||||
|
become aware of any circumstance which would make any of the foregoing
|
||||||
|
representations inaccurate in any respect. We may publicly disclose your
|
||||||
|
participation in the project, including the fact that you have signed the SCA.
|
||||||
|
|
||||||
|
6. This SCA is governed by the laws of the State of California and applicable
|
||||||
|
U.S. Federal law. Any choice of law rules will not apply.
|
||||||
|
|
||||||
|
7. Please place an “x” on one of the applicable statement below. Please do NOT
|
||||||
|
mark both statements:
|
||||||
|
|
||||||
|
* [X] I am signing on behalf of myself as an individual and no other person
|
||||||
|
or entity, including my employer, has or will have rights with respect to my
|
||||||
|
contributions.
|
||||||
|
|
||||||
|
* [ ] I am signing on behalf of my employer or a legal entity and I have the
|
||||||
|
actual authority to contractually bind that entity.
|
||||||
|
|
||||||
|
## Contributor Details
|
||||||
|
|
||||||
|
| Field | Entry |
|
||||||
|
|------------------------------- | -------------------- |
|
||||||
|
| Name | |
|
||||||
|
| Company name (if applicable) | |
|
||||||
|
| Title or role (if applicable) | |
|
||||||
|
| Date | |
|
||||||
|
| GitHub username | |
|
||||||
|
| Website (optional) | |
|
106
.github/contributors/pmbaumgartner.md
vendored
Normal file
106
.github/contributors/pmbaumgartner.md
vendored
Normal file
|
@ -0,0 +1,106 @@
|
||||||
|
# spaCy contributor agreement
|
||||||
|
|
||||||
|
This spaCy Contributor Agreement (**"SCA"**) is based on the
|
||||||
|
[Oracle Contributor Agreement](http://www.oracle.com/technetwork/oca-405177.pdf).
|
||||||
|
The SCA applies to any contribution that you make to any product or project
|
||||||
|
managed by us (the **"project"**), and sets out the intellectual property rights
|
||||||
|
you grant to us in the contributed materials. The term **"us"** shall mean
|
||||||
|
[ExplosionAI GmbH](https://explosion.ai/legal). The term
|
||||||
|
**"you"** shall mean the person or entity identified below.
|
||||||
|
|
||||||
|
If you agree to be bound by these terms, fill in the information requested
|
||||||
|
below and include the filled-in version with your first pull request, under the
|
||||||
|
folder [`.github/contributors/`](/.github/contributors/). The name of the file
|
||||||
|
should be your GitHub username, with the extension `.md`. For example, the user
|
||||||
|
example_user would create the file `.github/contributors/example_user.md`.
|
||||||
|
|
||||||
|
Read this agreement carefully before signing. These terms and conditions
|
||||||
|
constitute a binding legal agreement.
|
||||||
|
|
||||||
|
## Contributor Agreement
|
||||||
|
|
||||||
|
1. The term "contribution" or "contributed materials" means any source code,
|
||||||
|
object code, patch, tool, sample, graphic, specification, manual,
|
||||||
|
documentation, or any other material posted or submitted by you to the project.
|
||||||
|
|
||||||
|
2. With respect to any worldwide copyrights, or copyright applications and
|
||||||
|
registrations, in your contribution:
|
||||||
|
|
||||||
|
* you hereby assign to us joint ownership, and to the extent that such
|
||||||
|
assignment is or becomes invalid, ineffective or unenforceable, you hereby
|
||||||
|
grant to us a perpetual, irrevocable, non-exclusive, worldwide, no-charge,
|
||||||
|
royalty-free, unrestricted license to exercise all rights under those
|
||||||
|
copyrights. This includes, at our option, the right to sublicense these same
|
||||||
|
rights to third parties through multiple levels of sublicensees or other
|
||||||
|
licensing arrangements;
|
||||||
|
|
||||||
|
* you agree that each of us can do all things in relation to your
|
||||||
|
contribution as if each of us were the sole owners, and if one of us makes
|
||||||
|
a derivative work of your contribution, the one who makes the derivative
|
||||||
|
work (or has it made will be the sole owner of that derivative work;
|
||||||
|
|
||||||
|
* you agree that you will not assert any moral rights in your contribution
|
||||||
|
against us, our licensees or transferees;
|
||||||
|
|
||||||
|
* you agree that we may register a copyright in your contribution and
|
||||||
|
exercise all ownership rights associated with it; and
|
||||||
|
|
||||||
|
* you agree that neither of us has any duty to consult with, obtain the
|
||||||
|
consent of, pay or render an accounting to the other for any use or
|
||||||
|
distribution of your contribution.
|
||||||
|
|
||||||
|
3. With respect to any patents you own, or that you can license without payment
|
||||||
|
to any third party, you hereby grant to us a perpetual, irrevocable,
|
||||||
|
non-exclusive, worldwide, no-charge, royalty-free license to:
|
||||||
|
|
||||||
|
* make, have made, use, sell, offer to sell, import, and otherwise transfer
|
||||||
|
your contribution in whole or in part, alone or in combination with or
|
||||||
|
included in any product, work or materials arising out of the project to
|
||||||
|
which your contribution was submitted, and
|
||||||
|
|
||||||
|
* at our option, to sublicense these same rights to third parties through
|
||||||
|
multiple levels of sublicensees or other licensing arrangements.
|
||||||
|
|
||||||
|
4. Except as set out above, you keep all right, title, and interest in your
|
||||||
|
contribution. The rights that you grant to us under these terms are effective
|
||||||
|
on the date you first submitted a contribution to us, even if your submission
|
||||||
|
took place before the date you sign these terms.
|
||||||
|
|
||||||
|
5. You covenant, represent, warrant and agree that:
|
||||||
|
|
||||||
|
* Each contribution that you submit is and shall be an original work of
|
||||||
|
authorship and you can legally grant the rights set out in this SCA;
|
||||||
|
|
||||||
|
* to the best of your knowledge, each contribution will not violate any
|
||||||
|
third party's copyrights, trademarks, patents, or other intellectual
|
||||||
|
property rights; and
|
||||||
|
|
||||||
|
* each contribution shall be in compliance with U.S. export control laws and
|
||||||
|
other applicable export and import laws. You agree to notify us if you
|
||||||
|
become aware of any circumstance which would make any of the foregoing
|
||||||
|
representations inaccurate in any respect. We may publicly disclose your
|
||||||
|
participation in the project, including the fact that you have signed the SCA.
|
||||||
|
|
||||||
|
6. This SCA is governed by the laws of the State of California and applicable
|
||||||
|
U.S. Federal law. Any choice of law rules will not apply.
|
||||||
|
|
||||||
|
7. Please place an “x” on one of the applicable statement below. Please do NOT
|
||||||
|
mark both statements:
|
||||||
|
|
||||||
|
* [x] I am signing on behalf of myself as an individual and no other person
|
||||||
|
or entity, including my employer, has or will have rights with respect to my
|
||||||
|
contributions.
|
||||||
|
|
||||||
|
* [ ] I am signing on behalf of my employer or a legal entity and I have the
|
||||||
|
actual authority to contractually bind that entity.
|
||||||
|
|
||||||
|
## Contributor Details
|
||||||
|
|
||||||
|
| Field | Entry |
|
||||||
|
|------------------------------- | -------------------- |
|
||||||
|
| Name | Peter Baumgartner |
|
||||||
|
| Company name (if applicable) | |
|
||||||
|
| Title or role (if applicable) | |
|
||||||
|
| Date | July 14, 2019 |
|
||||||
|
| GitHub username | pmbaumgartner |
|
||||||
|
| Website (optional) | pmbaumgartner.github.io |
|
|
@ -67,7 +67,7 @@ valuable if it's shared publicly, so that more people can benefit from it.
|
||||||
|
|
||||||
- Non-destructive **tokenization**
|
- Non-destructive **tokenization**
|
||||||
- **Named entity** recognition
|
- **Named entity** recognition
|
||||||
- Support for **49+ languages**
|
- Support for **50+ languages**
|
||||||
- Pre-trained [statistical models](https://spacy.io/models) and word vectors
|
- Pre-trained [statistical models](https://spacy.io/models) and word vectors
|
||||||
- State-of-the-art speed
|
- State-of-the-art speed
|
||||||
- Easy **deep learning** integration
|
- Easy **deep learning** integration
|
||||||
|
|
|
@ -70,15 +70,33 @@ def merge_sents(sents):
|
||||||
return [(m_deps, m_brackets)]
|
return [(m_deps, m_brackets)]
|
||||||
|
|
||||||
|
|
||||||
def align(cand_words, gold_words):
|
def align(tokens_a, tokens_b):
|
||||||
if cand_words == gold_words:
|
"""Calculate alignment tables between two tokenizations, using the Levenshtein
|
||||||
alignment = numpy.arange(len(cand_words))
|
algorithm. The alignment is case-insensitive.
|
||||||
|
|
||||||
|
tokens_a (List[str]): The candidate tokenization.
|
||||||
|
tokens_b (List[str]): The reference tokenization.
|
||||||
|
RETURNS: (tuple): A 5-tuple consisting of the following information:
|
||||||
|
* cost (int): The number of misaligned tokens.
|
||||||
|
* a2b (List[int]): Mapping of indices in `tokens_a` to indices in `tokens_b`.
|
||||||
|
For instance, if `a2b[4] == 6`, that means that `tokens_a[4]` aligns
|
||||||
|
to `tokens_b[6]`. If there's no one-to-one alignment for a token,
|
||||||
|
it has the value -1.
|
||||||
|
* b2a (List[int]): The same as `a2b`, but mapping the other direction.
|
||||||
|
* a2b_multi (Dict[int, int]): A dictionary mapping indices in `tokens_a`
|
||||||
|
to indices in `tokens_b`, where multiple tokens of `tokens_a` align to
|
||||||
|
the same token of `tokens_b`.
|
||||||
|
* b2a_multi (Dict[int, int]): As with `a2b_multi`, but mapping the other
|
||||||
|
direction.
|
||||||
|
"""
|
||||||
|
if tokens_a == tokens_b:
|
||||||
|
alignment = numpy.arange(len(tokens_a))
|
||||||
return 0, alignment, alignment, {}, {}
|
return 0, alignment, alignment, {}, {}
|
||||||
cand_words = [w.replace(" ", "").lower() for w in cand_words]
|
tokens_a = [w.replace(" ", "").lower() for w in tokens_a]
|
||||||
gold_words = [w.replace(" ", "").lower() for w in gold_words]
|
tokens_b = [w.replace(" ", "").lower() for w in tokens_b]
|
||||||
cost, i2j, j2i, matrix = _align.align(cand_words, gold_words)
|
cost, i2j, j2i, matrix = _align.align(tokens_a, tokens_b)
|
||||||
i2j_multi, j2i_multi = _align.multi_align(i2j, j2i, [len(w) for w in cand_words],
|
i2j_multi, j2i_multi = _align.multi_align(i2j, j2i, [len(w) for w in tokens_a],
|
||||||
[len(w) for w in gold_words])
|
[len(w) for w in tokens_b])
|
||||||
for i, j in list(i2j_multi.items()):
|
for i, j in list(i2j_multi.items()):
|
||||||
if i2j_multi.get(i+1) != j and i2j_multi.get(i-1) != j:
|
if i2j_multi.get(i+1) != j and i2j_multi.get(i-1) != j:
|
||||||
i2j[i] = j
|
i2j[i] = j
|
||||||
|
|
|
@ -532,6 +532,7 @@ for orth in [
|
||||||
"Ms.",
|
"Ms.",
|
||||||
"p.m.",
|
"p.m.",
|
||||||
"Ph.D.",
|
"Ph.D.",
|
||||||
|
"Prof.",
|
||||||
"Rep.",
|
"Rep.",
|
||||||
"Rev.",
|
"Rev.",
|
||||||
"Sen.",
|
"Sen.",
|
||||||
|
|
|
@ -172,12 +172,8 @@ class Scorer(object):
|
||||||
self.ner_per_ents[ent.label_].fp += 1
|
self.ner_per_ents[ent.label_].fp += 1
|
||||||
else:
|
else:
|
||||||
cand_ents.add((ent.label_, first, last))
|
cand_ents.add((ent.label_, first, last))
|
||||||
current_ent[ent.label_].add(
|
current_ent[ent.label_].update([x for x in cand_ents if x[0] == ent.label_])
|
||||||
tuple(x for x in cand_ents if x[0] == ent.label_)
|
current_gold[ent.label_].update([x for x in gold_ents if x[0] == ent.label_])
|
||||||
)
|
|
||||||
current_gold[ent.label_].add(
|
|
||||||
tuple(x for x in gold_ents if x[0] == ent.label_)
|
|
||||||
)
|
|
||||||
# Scores per ent
|
# Scores per ent
|
||||||
[
|
[
|
||||||
v.score_set(current_ent[k], current_gold[k])
|
v.score_set(current_ent[k], current_gold[k])
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
# coding: utf8
|
# coding: utf8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import pytest
|
|
||||||
import spacy
|
import spacy
|
||||||
from spacy.util import minibatch, compounding
|
from spacy.util import minibatch, compounding
|
||||||
|
|
||||||
|
@ -9,27 +8,25 @@ from spacy.util import minibatch, compounding
|
||||||
def test_issue3611():
|
def test_issue3611():
|
||||||
""" Test whether adding n-grams in the textcat works even when n > token length of some docs """
|
""" Test whether adding n-grams in the textcat works even when n > token length of some docs """
|
||||||
unique_classes = ["offensive", "inoffensive"]
|
unique_classes = ["offensive", "inoffensive"]
|
||||||
x_train = ["This is an offensive text",
|
x_train = [
|
||||||
"This is the second offensive text",
|
"This is an offensive text",
|
||||||
"inoff"]
|
"This is the second offensive text",
|
||||||
|
"inoff",
|
||||||
|
]
|
||||||
y_train = ["offensive", "offensive", "inoffensive"]
|
y_train = ["offensive", "offensive", "inoffensive"]
|
||||||
|
|
||||||
# preparing the data
|
# preparing the data
|
||||||
pos_cats = list()
|
pos_cats = list()
|
||||||
for train_instance in y_train:
|
for train_instance in y_train:
|
||||||
pos_cats.append({label: label == train_instance for label in unique_classes})
|
pos_cats.append({label: label == train_instance for label in unique_classes})
|
||||||
train_data = list(zip(x_train, [{'cats': cats} for cats in pos_cats]))
|
train_data = list(zip(x_train, [{"cats": cats} for cats in pos_cats]))
|
||||||
|
|
||||||
# set up the spacy model with a text categorizer component
|
# set up the spacy model with a text categorizer component
|
||||||
nlp = spacy.blank('en')
|
nlp = spacy.blank("en")
|
||||||
|
|
||||||
textcat = nlp.create_pipe(
|
textcat = nlp.create_pipe(
|
||||||
"textcat",
|
"textcat",
|
||||||
config={
|
config={"exclusive_classes": True, "architecture": "bow", "ngram_size": 2},
|
||||||
"exclusive_classes": True,
|
|
||||||
"architecture": "bow",
|
|
||||||
"ngram_size": 2
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
|
|
||||||
for label in unique_classes:
|
for label in unique_classes:
|
||||||
|
@ -37,7 +34,7 @@ def test_issue3611():
|
||||||
nlp.add_pipe(textcat, last=True)
|
nlp.add_pipe(textcat, last=True)
|
||||||
|
|
||||||
# training the network
|
# training the network
|
||||||
other_pipes = [pipe for pipe in nlp.pipe_names if pipe != 'textcat']
|
other_pipes = [pipe for pipe in nlp.pipe_names if pipe != "textcat"]
|
||||||
with nlp.disable_pipes(*other_pipes):
|
with nlp.disable_pipes(*other_pipes):
|
||||||
optimizer = nlp.begin_training()
|
optimizer = nlp.begin_training()
|
||||||
for i in range(3):
|
for i in range(3):
|
||||||
|
@ -46,6 +43,10 @@ def test_issue3611():
|
||||||
|
|
||||||
for batch in batches:
|
for batch in batches:
|
||||||
texts, annotations = zip(*batch)
|
texts, annotations = zip(*batch)
|
||||||
nlp.update(docs=texts, golds=annotations, sgd=optimizer, drop=0.1, losses=losses)
|
nlp.update(
|
||||||
|
docs=texts,
|
||||||
|
golds=annotations,
|
||||||
|
sgd=optimizer,
|
||||||
|
drop=0.1,
|
||||||
|
losses=losses,
|
||||||
|
)
|
||||||
|
|
|
@ -3,8 +3,10 @@ from __future__ import unicode_literals
|
||||||
|
|
||||||
from spacy.lang.hi import Hindi
|
from spacy.lang.hi import Hindi
|
||||||
|
|
||||||
|
|
||||||
def test_issue3625():
|
def test_issue3625():
|
||||||
"""Test that default punctuation rules applies to hindi unicode characters"""
|
"""Test that default punctuation rules applies to hindi unicode characters"""
|
||||||
nlp = Hindi()
|
nlp = Hindi()
|
||||||
doc = nlp(u"hi. how हुए. होटल, होटल")
|
doc = nlp("hi. how हुए. होटल, होटल")
|
||||||
assert [token.text for token in doc] == ['hi', '.', 'how', 'हुए', '.', 'होटल', ',', 'होटल']
|
expected = ["hi", ".", "how", "हुए", ".", "होटल", ",", "होटल"]
|
||||||
|
assert [token.text for token in doc] == expected
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
# coding: utf8
|
# coding: utf8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import pytest
|
|
||||||
from spacy.matcher import Matcher
|
from spacy.matcher import Matcher
|
||||||
from spacy.tokens import Doc
|
from spacy.tokens import Doc
|
||||||
|
|
||||||
|
|
|
@ -2,7 +2,6 @@
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from spacy.attrs import IS_ALPHA
|
from spacy.attrs import IS_ALPHA
|
||||||
from spacy.lang.en import English
|
from spacy.lang.en import English
|
||||||
|
|
||||||
|
@ -10,11 +9,11 @@ from spacy.lang.en import English
|
||||||
@pytest.mark.parametrize(
|
@pytest.mark.parametrize(
|
||||||
"sentence",
|
"sentence",
|
||||||
[
|
[
|
||||||
'The story was to the effect that a young American student recently called on Professor Christlieb with a letter of introduction.',
|
"The story was to the effect that a young American student recently called on Professor Christlieb with a letter of introduction.",
|
||||||
'The next month Barry Siddall joined Stoke City on a free transfer, after Chris Pearce had established himself as the Vale\'s #1.',
|
"The next month Barry Siddall joined Stoke City on a free transfer, after Chris Pearce had established himself as the Vale's #1.",
|
||||||
'The next month Barry Siddall joined Stoke City on a free transfer, after Chris Pearce had established himself as the Vale\'s number one',
|
"The next month Barry Siddall joined Stoke City on a free transfer, after Chris Pearce had established himself as the Vale's number one",
|
||||||
'Indeed, making the one who remains do all the work has installed him into a position of such insolent tyranny, it will take a month at least to reduce him to his proper proportions.',
|
"Indeed, making the one who remains do all the work has installed him into a position of such insolent tyranny, it will take a month at least to reduce him to his proper proportions.",
|
||||||
"It was a missed assignment, but it shouldn't have resulted in a turnover ..."
|
"It was a missed assignment, but it shouldn't have resulted in a turnover ...",
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
def test_issue3869(sentence):
|
def test_issue3869(sentence):
|
||||||
|
@ -27,5 +26,3 @@ def test_issue3869(sentence):
|
||||||
count += token.is_alpha
|
count += token.is_alpha
|
||||||
|
|
||||||
assert count == doc.count_by(IS_ALPHA).get(1, 0)
|
assert count == doc.count_by(IS_ALPHA).get(1, 0)
|
||||||
|
|
||||||
|
|
||||||
|
|
22
spacy/tests/regression/test_issue3951.py
Normal file
22
spacy/tests/regression/test_issue3951.py
Normal file
|
@ -0,0 +1,22 @@
|
||||||
|
# coding: utf8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import pytest
|
||||||
|
from spacy.matcher import Matcher
|
||||||
|
from spacy.tokens import Doc
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.xfail
|
||||||
|
def test_issue3951(en_vocab):
|
||||||
|
"""Test that combinations of optional rules are matched correctly."""
|
||||||
|
matcher = Matcher(en_vocab)
|
||||||
|
pattern = [
|
||||||
|
{"LOWER": "hello"},
|
||||||
|
{"LOWER": "this", "OP": "?"},
|
||||||
|
{"OP": "?"},
|
||||||
|
{"LOWER": "world"},
|
||||||
|
]
|
||||||
|
matcher.add("TEST", None, pattern)
|
||||||
|
doc = Doc(en_vocab, words=["Hello", "my", "new", "world"])
|
||||||
|
matches = matcher(doc)
|
||||||
|
assert len(matches) == 0
|
34
spacy/tests/regression/test_issue3968.py
Normal file
34
spacy/tests/regression/test_issue3968.py
Normal file
|
@ -0,0 +1,34 @@
|
||||||
|
# coding: utf-8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
from spacy.gold import GoldParse
|
||||||
|
from spacy.scorer import Scorer
|
||||||
|
from ..util import get_doc
|
||||||
|
|
||||||
|
test_samples = [
|
||||||
|
[
|
||||||
|
"100 - 200",
|
||||||
|
{
|
||||||
|
"entities": [
|
||||||
|
[0, 3, "CARDINAL"],
|
||||||
|
[6, 9, "CARDINAL"]
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
]
|
||||||
|
|
||||||
|
def test_issue3625(en_vocab):
|
||||||
|
scorer = Scorer()
|
||||||
|
for input_, annot in test_samples:
|
||||||
|
doc = get_doc(en_vocab, words = input_.split(' '), ents = [[0,1,'CARDINAL'], [2,3,'CARDINAL']]);
|
||||||
|
gold = GoldParse(doc, entities = annot['entities'])
|
||||||
|
scorer.score(doc, gold)
|
||||||
|
results = scorer.scores
|
||||||
|
|
||||||
|
# Expects total accuracy and accuracy for each each entity to be 100%
|
||||||
|
assert results['ents_p'] == 100
|
||||||
|
assert results['ents_f'] == 100
|
||||||
|
assert results['ents_r'] == 100
|
||||||
|
assert results['ents_per_type']['CARDINAL']['p'] == 100
|
||||||
|
assert results['ents_per_type']['CARDINAL']['f'] == 100
|
||||||
|
assert results['ents_per_type']['CARDINAL']['r'] == 100
|
18
spacy/tests/regression/test_issue3972.py
Normal file
18
spacy/tests/regression/test_issue3972.py
Normal file
|
@ -0,0 +1,18 @@
|
||||||
|
# coding: utf8
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
|
import pytest
|
||||||
|
from spacy.matcher import PhraseMatcher
|
||||||
|
from spacy.tokens import Doc
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.xfail
|
||||||
|
def test_issue3972(en_vocab):
|
||||||
|
"""Test that the PhraseMatcher returns duplicates for duplicate match IDs.
|
||||||
|
"""
|
||||||
|
matcher = PhraseMatcher(en_vocab)
|
||||||
|
matcher.add("A", None, Doc(en_vocab, words=["New", "York"]))
|
||||||
|
matcher.add("B", None, Doc(en_vocab, words=["New", "York"]))
|
||||||
|
doc = Doc(en_vocab, words=["I", "live", "in", "New", "York"])
|
||||||
|
matches = matcher(doc)
|
||||||
|
assert len(matches) == 2
|
|
@ -209,12 +209,10 @@ set.
|
||||||
| `NE` | `PROPN` | | proper noun |
|
| `NE` | `PROPN` | | proper noun |
|
||||||
| `NNE` | `PROPN` | | proper noun |
|
| `NNE` | `PROPN` | | proper noun |
|
||||||
| `NN` | `NOUN` | | noun, singular or mass |
|
| `NN` | `NOUN` | | noun, singular or mass |
|
||||||
| `PAV` | `ADV` | `PronType=dem` | pronominal adverb |
|
|
||||||
| `PROAV` | `ADV` | `PronType=dem` | pronominal adverb |
|
| `PROAV` | `ADV` | `PronType=dem` | pronominal adverb |
|
||||||
| `PDAT` | `DET` | `PronType=dem` | attributive demonstrative pronoun |
|
| `PDAT` | `DET` | `PronType=dem` | attributive demonstrative pronoun |
|
||||||
| `PDS` | `PRON` | `PronType=dem` | substituting demonstrative pronoun |
|
| `PDS` | `PRON` | `PronType=dem` | substituting demonstrative pronoun |
|
||||||
| `PIAT` | `DET` | `PronType=ind\|neg\|tot` | attributive indefinite pronoun without determiner |
|
| `PIAT` | `DET` | `PronType=ind\|neg\|tot` | attributive indefinite pronoun without determiner |
|
||||||
| `PIDAT` | `DET` | `AdjType=pdt PronType=ind\|neg\|tot` | attributive indefinite pronoun with determiner |
|
|
||||||
| `PIS` | `PRON` | `PronType=ind\|neg\|tot` | substituting indefinite pronoun |
|
| `PIS` | `PRON` | `PronType=ind\|neg\|tot` | substituting indefinite pronoun |
|
||||||
| `PPER` | `PRON` | `PronType=prs` | non-reflexive personal pronoun |
|
| `PPER` | `PRON` | `PronType=prs` | non-reflexive personal pronoun |
|
||||||
| `PPOSAT` | `DET` | `Poss=yes PronType=prs` | attributive possessive pronoun |
|
| `PPOSAT` | `DET` | `Poss=yes PronType=prs` | attributive possessive pronoun |
|
||||||
|
@ -226,7 +224,7 @@ set.
|
||||||
| `PTKANT` | `PART` | `PartType=res` | answer particle |
|
| `PTKANT` | `PART` | `PartType=res` | answer particle |
|
||||||
| `PTKNEG` | `PART` | `Negative=yes` | negative particle |
|
| `PTKNEG` | `PART` | `Negative=yes` | negative particle |
|
||||||
| `PTKVZ` | `PART` | `PartType=vbp` | separable verbal particle |
|
| `PTKVZ` | `PART` | `PartType=vbp` | separable verbal particle |
|
||||||
| `PTKZU` | `PART` | `PartType=inf" | "zu" before infinitive |
|
| `PTKZU` | `PART` | `PartType=inf` | "zu" before infinitive |
|
||||||
| `PWAT` | `DET` | `PronType=int` | attributive interrogative pronoun |
|
| `PWAT` | `DET` | `PronType=int` | attributive interrogative pronoun |
|
||||||
| `PWAV` | `ADV` | `PronType=int` | adverbial interrogative or relative pronoun |
|
| `PWAV` | `ADV` | `PronType=int` | adverbial interrogative or relative pronoun |
|
||||||
| `PWS` | `PRON` | `PronType=int` | substituting interrogative pronoun |
|
| `PWS` | `PRON` | `PronType=int` | substituting interrogative pronoun |
|
||||||
|
@ -241,7 +239,7 @@ set.
|
||||||
| `VVFIN` | `VERB` | `Mood=ind VerbForm=fin` | finite verb, full |
|
| `VVFIN` | `VERB` | `Mood=ind VerbForm=fin` | finite verb, full |
|
||||||
| `VVIMP` | `VERB` | `Mood=imp VerbForm=fin` | imperative, full |
|
| `VVIMP` | `VERB` | `Mood=imp VerbForm=fin` | imperative, full |
|
||||||
| `VVINF` | `VERB` | `VerbForm=inf` | infinitive, full |
|
| `VVINF` | `VERB` | `VerbForm=inf` | infinitive, full |
|
||||||
| `VVIZU` | `VERB` | `VerbForm=inf" | infinitive with "zu", full |
|
| `VVIZU` | `VERB` | `VerbForm=inf` | infinitive with "zu", full |
|
||||||
| `VVPP` | `VERB` | `Aspect=perf VerbForm=part` | perfect participle, full |
|
| `VVPP` | `VERB` | `Aspect=perf VerbForm=part` | perfect participle, full |
|
||||||
| `XY` | `X` | | non-word containing non-letter |
|
| `XY` | `X` | | non-word containing non-letter |
|
||||||
| `SP` | `SPACE` | | space |
|
| `SP` | `SPACE` | | space |
|
||||||
|
@ -396,10 +394,8 @@ annotation scheme.
|
||||||
| `cp` | complementizer |
|
| `cp` | complementizer |
|
||||||
| `cvc` | collocational verb construction |
|
| `cvc` | collocational verb construction |
|
||||||
| `da` | dative |
|
| `da` | dative |
|
||||||
| `dh` | discourse-level head |
|
|
||||||
| `dm` | discourse marker |
|
| `dm` | discourse marker |
|
||||||
| `ep` | expletive es |
|
| `ep` | expletive es |
|
||||||
| `hd` | head |
|
|
||||||
| `ju` | junctor |
|
| `ju` | junctor |
|
||||||
| `mnr` | postnominal modifier |
|
| `mnr` | postnominal modifier |
|
||||||
| `mo` | modifier |
|
| `mo` | modifier |
|
||||||
|
@ -407,7 +403,7 @@ annotation scheme.
|
||||||
| `nk` | noun kernel element |
|
| `nk` | noun kernel element |
|
||||||
| `nmc` | numerical component |
|
| `nmc` | numerical component |
|
||||||
| `oa` | accusative object |
|
| `oa` | accusative object |
|
||||||
| `oa` | second accusative object |
|
| `oa2` | second accusative object |
|
||||||
| `oc` | clausal object |
|
| `oc` | clausal object |
|
||||||
| `og` | genitive object |
|
| `og` | genitive object |
|
||||||
| `op` | prepositional object |
|
| `op` | prepositional object |
|
||||||
|
@ -417,11 +413,13 @@ annotation scheme.
|
||||||
| `ph` | placeholder |
|
| `ph` | placeholder |
|
||||||
| `pm` | morphological particle |
|
| `pm` | morphological particle |
|
||||||
| `pnc` | proper noun component |
|
| `pnc` | proper noun component |
|
||||||
|
| `punct` | punctuation |
|
||||||
| `rc` | relative clause |
|
| `rc` | relative clause |
|
||||||
| `re` | repeated element |
|
| `re` | repeated element |
|
||||||
| `rs` | reported speech |
|
| `rs` | reported speech |
|
||||||
| `sb` | subject |
|
| `sb` | subject |
|
||||||
| `sp` | "subject or predicate |
|
| `sbp` | passivized subject (PP) |
|
||||||
|
| `sp` | subject or predicate |
|
||||||
| `svp` | separable verb prefix |
|
| `svp` | separable verb prefix |
|
||||||
| `uc` | unit component |
|
| `uc` | unit component |
|
||||||
| `vo` | vocative |
|
| `vo` | vocative |
|
||||||
|
|
|
@ -86,11 +86,11 @@ generate [Markdown](https://en.wikipedia.org/wiki/Markdown)-formatted markup to
|
||||||
copy-paste into [GitHub issues](https://github.com/explosion/spaCy/issues).
|
copy-paste into [GitHub issues](https://github.com/explosion/spaCy/issues).
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ python -m spacy info [--markdown]
|
$ python -m spacy info [--markdown] [--silent]
|
||||||
```
|
```
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
$ python -m spacy info [model] [--markdown]
|
$ python -m spacy info [model] [--markdown] [--silent]
|
||||||
```
|
```
|
||||||
|
|
||||||
| Argument | Type | Description |
|
| Argument | Type | Description |
|
||||||
|
@ -422,6 +422,7 @@ pip install dist/en_model-0.0.0.tar.gz
|
||||||
| `input_dir` | positional | Path to directory containing model data. |
|
| `input_dir` | positional | Path to directory containing model data. |
|
||||||
| `output_dir` | positional | Directory to create package folder in. |
|
| `output_dir` | positional | Directory to create package folder in. |
|
||||||
| `--meta-path`, `-m` <Tag variant="new">2</Tag> | option | Path to `meta.json` file (optional). |
|
| `--meta-path`, `-m` <Tag variant="new">2</Tag> | option | Path to `meta.json` file (optional). |
|
||||||
| `--create-meta`, `-c` <Tag variant="new">2</Tag> | flag | Create a `meta.json` file on the command line, even if one already exists in the directory. If an existing file is found, its entries will be shown as the defaults in the command line prompt. | `--force`, `-f` | flag | Force overwriting of existing folder in output directory. |
|
| `--create-meta`, `-c` <Tag variant="new">2</Tag> | flag | Create a `meta.json` file on the command line, even if one already exists in the directory. If an existing file is found, its entries will be shown as the defaults in the command line prompt.
|
||||||
|
| `--force`, `-f` | flag | Force overwriting of existing folder in output directory. |
|
||||||
| `--help`, `-h` | flag | Show help message and available arguments. |
|
| `--help`, `-h` | flag | Show help message and available arguments. |
|
||||||
| **CREATES** | directory | A Python package containing the spaCy model. |
|
| **CREATES** | directory | A Python package containing the spaCy model. |
|
||||||
|
|
|
@ -23,7 +23,7 @@ be a token pattern (list) or a phrase pattern (string). For example:
|
||||||
>
|
>
|
||||||
> ```python
|
> ```python
|
||||||
> # Construction via create_pipe
|
> # Construction via create_pipe
|
||||||
> ruler = nlp.create_pipe("entityruler")
|
> ruler = nlp.create_pipe("entity_ruler")
|
||||||
>
|
>
|
||||||
> # Construction from class
|
> # Construction from class
|
||||||
> from spacy.pipeline import EntityRuler
|
> from spacy.pipeline import EntityRuler
|
||||||
|
|
|
@ -76,6 +76,59 @@ Convert a list of Doc objects into the
|
||||||
| `id` | int | ID to assign to the JSON. Defaults to `0`. |
|
| `id` | int | ID to assign to the JSON. Defaults to `0`. |
|
||||||
| **RETURNS** | list | The data in spaCy's JSON format. |
|
| **RETURNS** | list | The data in spaCy's JSON format. |
|
||||||
|
|
||||||
|
### gold.align {#align tag="function"}
|
||||||
|
|
||||||
|
Calculate alignment tables between two tokenizations, using the Levenshtein
|
||||||
|
algorithm. The alignment is case-insensitive.
|
||||||
|
|
||||||
|
<Infobox title="Important note" variant="warning">
|
||||||
|
|
||||||
|
The current implementation of the alignment algorithm assumes that both
|
||||||
|
tokenizations add up to the same string. For example, you'll be able to align
|
||||||
|
`["I", "'", "m"]` and `["I", "'m"]`, which both add up to `"I'm"`, but not
|
||||||
|
`["I", "'m"]` and `["I", "am"]`.
|
||||||
|
|
||||||
|
</Infobox>
|
||||||
|
|
||||||
|
> #### Example
|
||||||
|
>
|
||||||
|
> ```python
|
||||||
|
> from spacy.gold import align
|
||||||
|
>
|
||||||
|
> bert_tokens = ["obama", "'", "s", "podcast"]
|
||||||
|
> spacy_tokens = ["obama", "'s", "podcast"]
|
||||||
|
> alignment = align(bert_tokens, spacy_tokens)
|
||||||
|
> cost, a2b, b2a, a2b_multi, b2a_multi = alignment
|
||||||
|
> ```
|
||||||
|
|
||||||
|
| Name | Type | Description |
|
||||||
|
| ----------- | ----- | -------------------------------------------------------------------------- |
|
||||||
|
| `tokens_a` | list | String values of candidate tokens to align. |
|
||||||
|
| `tokens_b` | list | String values of reference tokens to align. |
|
||||||
|
| **RETURNS** | tuple | A `(cost, a2b, b2a, a2b_multi, b2a_multi)` tuple describing the alignment. |
|
||||||
|
|
||||||
|
The returned tuple contains the following alignment information:
|
||||||
|
|
||||||
|
> #### Example
|
||||||
|
>
|
||||||
|
> ```python
|
||||||
|
> a2b = array([0, -1, -1, 2])
|
||||||
|
> b2a = array([0, 2, 3])
|
||||||
|
> a2b_multi = {1: 1, 2: 1}
|
||||||
|
> b2a_multi = {}
|
||||||
|
> ```
|
||||||
|
>
|
||||||
|
> If `a2b[3] == 2`, that means that `tokens_a[3]` aligns to `tokens_b[2]`. If
|
||||||
|
> there's no one-to-one alignment for a token, it has the value `-1`.
|
||||||
|
|
||||||
|
| Name | Type | Description |
|
||||||
|
| ----------- | -------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------------------- |
|
||||||
|
| `cost` | int | The number of misaligned tokens. |
|
||||||
|
| `a2b` | `numpy.ndarray[ndim=1, dtype='int32']` | One-to-one mappings of indices in `tokens_a` to indices in `tokens_b`. |
|
||||||
|
| `b2a` | `numpy.ndarray[ndim=1, dtype='int32']` | One-to-one mappings of indices in `tokens_b` to indices in `tokens_a`. |
|
||||||
|
| `a2b_multi` | dict | A dictionary mapping indices in `tokens_a` to indices in `tokens_b`, where multiple tokens of `tokens_a` align to the same token of `tokens_b`. |
|
||||||
|
| `b2a_multi` | dict | A dictionary mapping indices in `tokens_b` to indices in `tokens_a`, where multiple tokens of `tokens_b` align to the same token of `tokens_a`. |
|
||||||
|
|
||||||
### gold.biluo_tags_from_offsets {#biluo_tags_from_offsets tag="function"}
|
### gold.biluo_tags_from_offsets {#biluo_tags_from_offsets tag="function"}
|
||||||
|
|
||||||
Encode labelled spans into per-token tags, using the
|
Encode labelled spans into per-token tags, using the
|
||||||
|
|
|
@ -963,6 +963,72 @@ Once you have a [`Doc`](/api/doc) object, you can write to its attributes to set
|
||||||
the part-of-speech tags, syntactic dependencies, named entities and other
|
the part-of-speech tags, syntactic dependencies, named entities and other
|
||||||
attributes. For details, see the respective usage pages.
|
attributes. For details, see the respective usage pages.
|
||||||
|
|
||||||
|
### Aligning tokenization {#aligning-tokenization}
|
||||||
|
|
||||||
|
spaCy's tokenization is non-destructive and uses language-specific rules
|
||||||
|
optimized for compatibility with treebank annotations. Other tools and resources
|
||||||
|
can sometimes tokenize things differently – for example, `"I'm"` →
|
||||||
|
`["I", "'", "m"]` instead of `["I", "'m"]`.
|
||||||
|
|
||||||
|
In situations like that, you often want to align the tokenization so that you
|
||||||
|
can merge annotations from different sources together, or take vectors predicted
|
||||||
|
by a
|
||||||
|
[pre-trained BERT model](https://github.com/huggingface/pytorch-transformers)
|
||||||
|
and apply them to spaCy tokens. spaCy's [`gold.align`](/api/goldparse#align)
|
||||||
|
helper returns a `(cost, a2b, b2a, a2b_multi, b2a_multi)` tuple describing the
|
||||||
|
number of misaligned tokens, the one-to-one mappings of token indices in both
|
||||||
|
directions and the indices where multiple tokens align to one single token.
|
||||||
|
|
||||||
|
> #### ✏️ Things to try
|
||||||
|
>
|
||||||
|
> 1. Change the capitalization in one of the token lists – for example,
|
||||||
|
> `"obama"` to `"Obama"`. You'll see that the alignment is case-insensitive.
|
||||||
|
> 2. Change `"podcasts"` in `other_tokens` to `"pod", "casts"`. You should see
|
||||||
|
> that there are now 4 misaligned tokens and that the new many-to-one mapping
|
||||||
|
> is reflected in `a2b_multi`.
|
||||||
|
> 3. Make `other_tokens` and `spacy_tokens` identical. You'll see that the
|
||||||
|
> `cost` is `0` and all corresponding mappings are also identical.
|
||||||
|
|
||||||
|
```python
|
||||||
|
### {executable="true"}
|
||||||
|
from spacy.gold import align
|
||||||
|
|
||||||
|
other_tokens = ["i", "listened", "to", "obama", "'", "s", "podcasts", "."]
|
||||||
|
spacy_tokens = ["i", "listened", "to", "obama", "'s", "podcasts", "."]
|
||||||
|
cost, a2b, b2a, a2b_multi, b2a_multi = align(other_tokens, spacy_tokens)
|
||||||
|
print("Misaligned tokens:", cost) # 2
|
||||||
|
print("One-to-one mappings a -> b", a2b) # array([0, 1, 2, 3, -1, -1, 5, 6])
|
||||||
|
print("One-to-one mappings b -> a", b2a) # array([0, 1, 2, 3, 5, 6, 7])
|
||||||
|
print("Many-to-one mappings a -> b", a2b_multi) # {4: 4, 5: 4}
|
||||||
|
print("Many-to-one mappings b-> a", b2a_multi) # {}
|
||||||
|
```
|
||||||
|
|
||||||
|
Here are some insights from the alignment information generated in the example
|
||||||
|
above:
|
||||||
|
|
||||||
|
- Two tokens are misaligned.
|
||||||
|
- The one-to-one mappings for the first four tokens are identical, which means
|
||||||
|
they map to each other. This makes sense because they're also identical in the
|
||||||
|
input: `"i"`, `"listened"`, `"to"` and `"obama"`.
|
||||||
|
- The index mapped to `a2b[6]` is `5`, which means that `other_tokens[6]`
|
||||||
|
(`"podcasts"`) aligns to `spacy_tokens[5]` (also `"podcasts"`).
|
||||||
|
- `a2b[4]` is `-1`, which means that there is no one-to-one alignment for the
|
||||||
|
token at `other_tokens[4]`. The token `"'"` doesn't exist on its own in
|
||||||
|
`spacy_tokens`. The same goes for `a2b[5]` and `other_tokens[5]`, i.e. `"s"`.
|
||||||
|
- The dictionary `a2b_multi` shows that both tokens 4 and 5 of `other_tokens`
|
||||||
|
(`"'"` and `"s"`) align to token 4 of `spacy_tokens` (`"'s"`).
|
||||||
|
- The dictionary `b2a_multi` shows that there are no tokens in `spacy_tokens`
|
||||||
|
that map to multiple tokens in `other_tokens`.
|
||||||
|
|
||||||
|
<Infobox title="Important note" variant="warning">
|
||||||
|
|
||||||
|
The current implementation of the alignment algorithm assumes that both
|
||||||
|
tokenizations add up to the same string. For example, you'll be able to align
|
||||||
|
`["I", "'", "m"]` and `["I", "'m"]`, which both add up to `"I'm"`, but not
|
||||||
|
`["I", "'m"]` and `["I", "am"]`.
|
||||||
|
|
||||||
|
</Infobox>
|
||||||
|
|
||||||
## Merging and splitting {#retokenization new="2.1"}
|
## Merging and splitting {#retokenization new="2.1"}
|
||||||
|
|
||||||
The [`Doc.retokenize`](/api/doc#retokenize) context manager lets you merge and
|
The [`Doc.retokenize`](/api/doc#retokenize) context manager lets you merge and
|
||||||
|
|
|
@ -169,7 +169,7 @@ rule-based matching are:
|
||||||
| `LIKE_NUM`, `LIKE_URL`, `LIKE_EMAIL` | bool | Token text resembles a number, URL, email. |
|
| `LIKE_NUM`, `LIKE_URL`, `LIKE_EMAIL` | bool | Token text resembles a number, URL, email. |
|
||||||
| `POS`, `TAG`, `DEP`, `LEMMA`, `SHAPE` | unicode | The token's simple and extended part-of-speech tag, dependency label, lemma, shape. |
|
| `POS`, `TAG`, `DEP`, `LEMMA`, `SHAPE` | unicode | The token's simple and extended part-of-speech tag, dependency label, lemma, shape. |
|
||||||
| `ENT_TYPE` | unicode | The token's entity label. |
|
| `ENT_TYPE` | unicode | The token's entity label. |
|
||||||
| `_` <Tag variant="new">2.1</Tag> | dict | Properties in [custom extension attributes](/processing-pipelines#custom-components-attributes). |
|
| `_` <Tag variant="new">2.1</Tag> | dict | Properties in [custom extension attributes](/usage/processing-pipelines#custom-components-attributes). |
|
||||||
|
|
||||||
<Infobox title="Tip: Try the interactive matcher explorer">
|
<Infobox title="Tip: Try the interactive matcher explorer">
|
||||||
|
|
||||||
|
@ -1140,7 +1140,7 @@ To apply this logic automatically when we process a text, we can add it to the
|
||||||
above logic also expects that entities are merged into single tokens. spaCy
|
above logic also expects that entities are merged into single tokens. spaCy
|
||||||
ships with a handy built-in `merge_entities` that takes care of that. Instead of
|
ships with a handy built-in `merge_entities` that takes care of that. Instead of
|
||||||
just printing the result, you could also write it to
|
just printing the result, you could also write it to
|
||||||
[custom attributes](/processing-pipelines#custom-components-attributes) on the
|
[custom attributes](/usage/processing-pipelines#custom-components-attributes) on the
|
||||||
entity `Span` – for example `._.orgs` or `._.prev_orgs` and `._.current_orgs`.
|
entity `Span` – for example `._.orgs` or `._.prev_orgs` and `._.current_orgs`.
|
||||||
|
|
||||||
> #### Merging entities
|
> #### Merging entities
|
||||||
|
|
Loading…
Reference in New Issue
Block a user