mirror of
https://github.com/explosion/spaCy.git
synced 2024-11-11 12:18:04 +03:00
db55577c45
* Remove unicode declarations * Remove Python 3.5 and 2.7 from CI * Don't require pathlib * Replace compat helpers * Remove OrderedDict * Use f-strings * Set Cython compiler language level * Fix typo * Re-add OrderedDict for Table * Update setup.cfg * Revert CONTRIBUTING.md * Revert lookups.md * Revert top-level.md * Small adjustments and docs [ci skip]
126 lines
4.0 KiB
Python
126 lines
4.0 KiB
Python
import re
|
|
from collections import namedtuple
|
|
|
|
from .stop_words import STOP_WORDS
|
|
from .tag_map import TAG_MAP
|
|
from ...attrs import LANG
|
|
from ...language import Language
|
|
from ...tokens import Doc
|
|
from ...compat import copy_reg
|
|
from ...util import DummyTokenizer
|
|
|
|
# Handling for multiple spaces in a row is somewhat awkward, this simplifies
|
|
# the flow by creating a dummy with the same interface.
|
|
DummyNode = namedtuple("DummyNode", ["surface", "pos", "feature"])
|
|
DummyNodeFeatures = namedtuple("DummyNodeFeatures", ["lemma"])
|
|
DummySpace = DummyNode(" ", " ", DummyNodeFeatures(" "))
|
|
|
|
|
|
def try_fugashi_import():
|
|
"""Fugashi is required for Japanese support, so check for it.
|
|
It it's not available blow up and explain how to fix it."""
|
|
try:
|
|
import fugashi
|
|
|
|
return fugashi
|
|
except ImportError:
|
|
raise ImportError(
|
|
"Japanese support requires Fugashi: " "https://github.com/polm/fugashi"
|
|
)
|
|
|
|
|
|
def resolve_pos(token):
|
|
"""If necessary, add a field to the POS tag for UD mapping.
|
|
Under Universal Dependencies, sometimes the same Unidic POS tag can
|
|
be mapped differently depending on the literal token or its context
|
|
in the sentence. This function adds information to the POS tag to
|
|
resolve ambiguous mappings.
|
|
"""
|
|
|
|
# this is only used for consecutive ascii spaces
|
|
if token.surface == " ":
|
|
return "空白"
|
|
|
|
# TODO: This is a first take. The rules here are crude approximations.
|
|
# For many of these, full dependencies are needed to properly resolve
|
|
# PoS mappings.
|
|
if token.pos == "連体詞,*,*,*":
|
|
if re.match(r"[こそあど此其彼]の", token.surface):
|
|
return token.pos + ",DET"
|
|
if re.match(r"[こそあど此其彼]", token.surface):
|
|
return token.pos + ",PRON"
|
|
return token.pos + ",ADJ"
|
|
return token.pos
|
|
|
|
|
|
def get_words_and_spaces(tokenizer, text):
|
|
"""Get the individual tokens that make up the sentence and handle white space.
|
|
|
|
Japanese doesn't usually use white space, and MeCab's handling of it for
|
|
multiple spaces in a row is somewhat awkward.
|
|
"""
|
|
|
|
tokens = tokenizer.parseToNodeList(text)
|
|
|
|
words = []
|
|
spaces = []
|
|
for token in tokens:
|
|
# If there's more than one space, spaces after the first become tokens
|
|
for ii in range(len(token.white_space) - 1):
|
|
words.append(DummySpace)
|
|
spaces.append(False)
|
|
|
|
words.append(token)
|
|
spaces.append(bool(token.white_space))
|
|
return words, spaces
|
|
|
|
|
|
class JapaneseTokenizer(DummyTokenizer):
|
|
def __init__(self, cls, nlp=None):
|
|
self.vocab = nlp.vocab if nlp is not None else cls.create_vocab(nlp)
|
|
self.tokenizer = try_fugashi_import().Tagger()
|
|
self.tokenizer.parseToNodeList("") # see #2901
|
|
|
|
def __call__(self, text):
|
|
dtokens, spaces = get_words_and_spaces(self.tokenizer, text)
|
|
words = [x.surface for x in dtokens]
|
|
doc = Doc(self.vocab, words=words, spaces=spaces)
|
|
unidic_tags = []
|
|
for token, dtoken in zip(doc, dtokens):
|
|
unidic_tags.append(dtoken.pos)
|
|
token.tag_ = resolve_pos(dtoken)
|
|
|
|
# if there's no lemma info (it's an unk) just use the surface
|
|
token.lemma_ = dtoken.feature.lemma or dtoken.surface
|
|
doc.user_data["unidic_tags"] = unidic_tags
|
|
return doc
|
|
|
|
|
|
class JapaneseDefaults(Language.Defaults):
|
|
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
|
|
lex_attr_getters[LANG] = lambda _text: "ja"
|
|
stop_words = STOP_WORDS
|
|
tag_map = TAG_MAP
|
|
writing_system = {"direction": "ltr", "has_case": False, "has_letters": False}
|
|
|
|
@classmethod
|
|
def create_tokenizer(cls, nlp=None):
|
|
return JapaneseTokenizer(cls, nlp)
|
|
|
|
|
|
class Japanese(Language):
|
|
lang = "ja"
|
|
Defaults = JapaneseDefaults
|
|
|
|
def make_doc(self, text):
|
|
return self.tokenizer(text)
|
|
|
|
|
|
def pickle_japanese(instance):
|
|
return Japanese, tuple()
|
|
|
|
|
|
copy_reg.pickle(Japanese, pickle_japanese)
|
|
|
|
__all__ = ["Japanese"]
|