mirror of
https://github.com/explosion/spaCy.git
synced 2024-12-24 17:06:29 +03:00
Fix circular import
This commit is contained in:
commit
1289187279
|
@ -3,12 +3,15 @@ from __future__ import unicode_literals
|
||||||
|
|
||||||
from .cli.info import info as cli_info
|
from .cli.info import info as cli_info
|
||||||
from .glossary import explain
|
from .glossary import explain
|
||||||
|
<<<<<<< HEAD
|
||||||
from .deprecated import resolve_load_name
|
from .deprecated import resolve_load_name
|
||||||
#from .about import __version__
|
#from .about import __version__
|
||||||
|
from .about import __version__
|
||||||
from . import util
|
from . import util
|
||||||
|
|
||||||
|
|
||||||
def load(name, **overrides):
|
def load(name, **overrides):
|
||||||
|
from .deprecated import resolve_load_name
|
||||||
name = resolve_load_name(name, **overrides)
|
name = resolve_load_name(name, **overrides)
|
||||||
return util.load_model(name, **overrides)
|
return util.load_model(name, **overrides)
|
||||||
|
|
||||||
|
|
|
@ -14,7 +14,7 @@ from ..util import prints
|
||||||
CONVERTERS = {
|
CONVERTERS = {
|
||||||
'.conllu': conllu2json,
|
'.conllu': conllu2json,
|
||||||
'.conll': conllu2json,
|
'.conll': conllu2json,
|
||||||
'.iob': iob2json
|
'.iob': iob2json,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,5 +1,6 @@
|
||||||
# coding: utf8
|
# coding: utf8
|
||||||
from __future__ import unicode_literals
|
from __future__ import unicode_literals
|
||||||
|
from cytoolz import partition_all, concat
|
||||||
|
|
||||||
from ...compat import json_dumps, path2str
|
from ...compat import json_dumps, path2str
|
||||||
from ...util import prints
|
from ...util import prints
|
||||||
|
@ -10,11 +11,9 @@ def iob2json(input_path, output_path, n_sents=10, *a, **k):
|
||||||
"""
|
"""
|
||||||
Convert IOB files into JSON format for use with train cli.
|
Convert IOB files into JSON format for use with train cli.
|
||||||
"""
|
"""
|
||||||
# TODO: This isn't complete yet -- need to map from IOB to
|
|
||||||
# BILUO
|
|
||||||
with input_path.open('r', encoding='utf8') as file_:
|
with input_path.open('r', encoding='utf8') as file_:
|
||||||
docs = read_iob(file_)
|
sentences = read_iob(file_)
|
||||||
|
docs = merge_sentences(sentences, n_sents)
|
||||||
output_filename = input_path.parts[-1].replace(".iob", ".json")
|
output_filename = input_path.parts[-1].replace(".iob", ".json")
|
||||||
output_file = output_path / output_filename
|
output_file = output_path / output_filename
|
||||||
with output_file.open('w', encoding='utf-8') as f:
|
with output_file.open('w', encoding='utf-8') as f:
|
||||||
|
@ -23,9 +22,9 @@ def iob2json(input_path, output_path, n_sents=10, *a, **k):
|
||||||
title="Generated output file %s" % path2str(output_file))
|
title="Generated output file %s" % path2str(output_file))
|
||||||
|
|
||||||
|
|
||||||
def read_iob(file_):
|
def read_iob(raw_sents):
|
||||||
sentences = []
|
sentences = []
|
||||||
for line in file_:
|
for line in raw_sents:
|
||||||
if not line.strip():
|
if not line.strip():
|
||||||
continue
|
continue
|
||||||
tokens = [t.split('|') for t in line.split()]
|
tokens = [t.split('|') for t in line.split()]
|
||||||
|
@ -43,3 +42,15 @@ def read_iob(file_):
|
||||||
paragraphs = [{'sentences': [sent]} for sent in sentences]
|
paragraphs = [{'sentences': [sent]} for sent in sentences]
|
||||||
docs = [{'id': 0, 'paragraphs': [para]} for para in paragraphs]
|
docs = [{'id': 0, 'paragraphs': [para]} for para in paragraphs]
|
||||||
return docs
|
return docs
|
||||||
|
|
||||||
|
def merge_sentences(docs, n_sents):
|
||||||
|
counter = 0
|
||||||
|
merged = []
|
||||||
|
for group in partition_all(n_sents, docs):
|
||||||
|
group = list(group)
|
||||||
|
first = group.pop(0)
|
||||||
|
to_extend = first['paragraphs'][0]['sentences']
|
||||||
|
for sent in group[1:]:
|
||||||
|
to_extend.extend(sent['paragraphs'][0]['sentences'])
|
||||||
|
merged.append(first)
|
||||||
|
return merged
|
||||||
|
|
|
@ -69,10 +69,10 @@ def train(cmd, lang, output_dir, train_data, dev_data, n_iter=10, n_sents=0,
|
||||||
prints("Expected dict but got: {}".format(type(meta)),
|
prints("Expected dict but got: {}".format(type(meta)),
|
||||||
title="Not a valid meta.json format", exits=1)
|
title="Not a valid meta.json format", exits=1)
|
||||||
|
|
||||||
pipeline = ['tags', 'dependencies', 'entities']
|
pipeline = ['tagger', 'parser', 'ner']
|
||||||
if no_tagger and 'tags' in pipeline: pipeline.remove('tags')
|
if no_tagger and 'tagger' in pipeline: pipeline.remove('tagger')
|
||||||
if no_parser and 'dependencies' in pipeline: pipeline.remove('dependencies')
|
if no_parser and 'parser' in pipeline: pipeline.remove('parser')
|
||||||
if no_entities and 'entities' in pipeline: pipeline.remove('entities')
|
if no_entities and 'ner' in pipeline: pipeline.remove('ner')
|
||||||
|
|
||||||
# Take dropout and batch size as generators of values -- dropout
|
# Take dropout and batch size as generators of values -- dropout
|
||||||
# starts high and decays sharply, to force the optimizer to explore.
|
# starts high and decays sharply, to force the optimizer to explore.
|
||||||
|
|
Loading…
Reference in New Issue
Block a user