2020-06-21 14:44:00 +03:00
|
|
|
from typing import Optional
|
|
|
|
from enum import Enum
|
2017-04-13 14:51:54 +03:00
|
|
|
from pathlib import Path
|
2018-11-30 22:16:14 +03:00
|
|
|
from wasabi import Printer
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 03:28:22 +03:00
|
|
|
import srsly
|
2019-08-29 13:04:01 +03:00
|
|
|
import re
|
2017-04-07 14:04:17 +03:00
|
|
|
|
2020-06-21 14:44:00 +03:00
|
|
|
from ._app import app, Arg, Opt
|
2019-03-15 20:14:46 +03:00
|
|
|
from .converters import conllu2json, iob2json, conll_ner2json
|
2018-08-14 15:04:32 +03:00
|
|
|
from .converters import ner_jsonl2json
|
2018-11-30 22:16:14 +03:00
|
|
|
|
2017-04-07 14:04:17 +03:00
|
|
|
|
2019-08-29 13:04:01 +03:00
|
|
|
# Converters are matched by file extension except for ner/iob, which are
|
|
|
|
# matched by file extension and content. To add a converter, add a new
|
2017-10-27 15:38:39 +03:00
|
|
|
# entry to this dict with the file extension mapped to the converter function
|
|
|
|
# imported from /converters.
|
2017-04-07 14:04:17 +03:00
|
|
|
CONVERTERS = {
|
2019-03-15 20:14:46 +03:00
|
|
|
"conllubio": conllu2json,
|
2018-11-30 22:16:14 +03:00
|
|
|
"conllu": conllu2json,
|
|
|
|
"conll": conllu2json,
|
|
|
|
"ner": conll_ner2json,
|
|
|
|
"iob": iob2json,
|
|
|
|
"jsonl": ner_jsonl2json,
|
2017-04-07 14:04:17 +03:00
|
|
|
}
|
|
|
|
|
2018-11-30 22:16:14 +03:00
|
|
|
# File types
|
2019-03-09 01:15:23 +03:00
|
|
|
FILE_TYPES_STDOUT = ("json", "jsonl")
|
2018-11-30 22:16:14 +03:00
|
|
|
|
2017-04-07 14:04:17 +03:00
|
|
|
|
2020-06-21 14:44:00 +03:00
|
|
|
class FileTypes(str, Enum):
|
|
|
|
json = "json"
|
|
|
|
jsonl = "jsonl"
|
|
|
|
msg = "msg"
|
|
|
|
|
|
|
|
|
|
|
|
@app.command("convert")
|
2020-06-21 22:35:01 +03:00
|
|
|
def convert_cli(
|
2019-12-22 03:53:56 +03:00
|
|
|
# fmt: off
|
2020-06-21 22:35:01 +03:00
|
|
|
input_file: str = Arg(..., help="Input file", exists=True),
|
|
|
|
output_dir: Path = Arg("-", help="Output directory. '-' for stdout.", allow_dash=True, exists=True),
|
2020-06-21 14:44:00 +03:00
|
|
|
file_type: FileTypes = Opt(FileTypes.json.value, "--file-type", "-t", help="Type of data to produce"),
|
|
|
|
n_sents: int = Opt(1, "--n-sents", "-n", help="Number of sentences per doc (0 to disable)"),
|
|
|
|
seg_sents: bool = Opt(False, "--seg-sents", "-s", help="Segment sentences (for -c ner)"),
|
|
|
|
model: Optional[str] = Opt(None, "--model", "-b", help="Model for sentence segmentation (for -s)"),
|
|
|
|
morphology: bool = Opt(False, "--morphology", "-m", help="Enable appending morphology to tags"),
|
|
|
|
merge_subtokens: bool = Opt(False, "--merge-subtokens", "-T", help="Merge CoNLL-U subtokens"),
|
|
|
|
converter: str = Opt("auto", "--converter", "-c", help=f"Converter: {tuple(CONVERTERS.keys())}"),
|
2020-06-21 22:35:01 +03:00
|
|
|
ner_map_path: Optional[Path] = Opt(None, "--ner-map-path", "-N", help="NER tag mapping (as JSON-encoded dict of entity types)", exists=True),
|
2020-06-21 14:44:00 +03:00
|
|
|
lang: Optional[str] = Opt(None, "--lang", "-l", help="Language (if tokenizer required)"),
|
2019-12-22 03:53:56 +03:00
|
|
|
# fmt: on
|
2018-11-30 22:16:14 +03:00
|
|
|
):
|
2017-05-27 21:01:46 +03:00
|
|
|
"""
|
|
|
|
Convert files into JSON format for use with train command and other
|
2018-11-30 22:16:14 +03:00
|
|
|
experiment management functions. If no output_dir is specified, the data
|
2019-04-12 12:31:23 +03:00
|
|
|
is written to stdout, so you can pipe them forward to a JSON file:
|
|
|
|
$ spacy convert some_file.conllu > some_file.json
|
2017-05-22 13:28:58 +03:00
|
|
|
"""
|
2020-06-21 14:44:00 +03:00
|
|
|
if isinstance(file_type, FileTypes):
|
|
|
|
# We get an instance of the FileTypes from the CLI so we need its string value
|
|
|
|
file_type = file_type.value
|
2020-06-21 22:35:01 +03:00
|
|
|
silent = output_dir == "-"
|
|
|
|
convert(
|
|
|
|
input_file,
|
|
|
|
output_dir,
|
|
|
|
file_type=file_type,
|
|
|
|
n_sents=n_sents,
|
|
|
|
seg_sents=seg_sents,
|
|
|
|
model=model,
|
|
|
|
morphology=morphology,
|
|
|
|
merge_subtokens=merge_subtokens,
|
|
|
|
converter=converter,
|
|
|
|
ner_map_path=ner_map_path,
|
|
|
|
lang=lang,
|
|
|
|
silent=silent,
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
def convert(
|
|
|
|
input_file: Path,
|
|
|
|
output_dir: Path,
|
|
|
|
*,
|
|
|
|
file_type: str = "json",
|
|
|
|
n_sents: int = 1,
|
|
|
|
seg_sents: bool = False,
|
|
|
|
model: Optional[str] = None,
|
|
|
|
morphology: bool = False,
|
|
|
|
merge_subtokens: bool = False,
|
|
|
|
converter: str = "auto",
|
|
|
|
ner_map_path: Optional[Path] = None,
|
|
|
|
lang: Optional[str] = None,
|
|
|
|
silent: bool = True,
|
|
|
|
) -> None:
|
|
|
|
msg = Printer(no_print=silent, pretty=not silent)
|
2017-04-07 14:04:17 +03:00
|
|
|
input_path = Path(input_file)
|
2019-03-09 01:15:23 +03:00
|
|
|
if file_type not in FILE_TYPES_STDOUT and output_dir == "-":
|
|
|
|
# TODO: support msgpack via stdout in srsly?
|
|
|
|
msg.fail(
|
2019-12-22 03:53:56 +03:00
|
|
|
f"Can't write .{file_type} data to stdout",
|
2019-03-09 01:15:23 +03:00
|
|
|
"Please specify an output directory.",
|
|
|
|
exits=1,
|
|
|
|
)
|
2017-05-08 00:25:29 +03:00
|
|
|
if not input_path.exists():
|
2018-12-08 13:49:43 +03:00
|
|
|
msg.fail("Input file not found", input_path, exits=1)
|
2018-11-30 22:16:14 +03:00
|
|
|
if output_dir != "-" and not Path(output_dir).exists():
|
2018-12-08 13:49:43 +03:00
|
|
|
msg.fail("Output directory not found", output_dir, exits=1)
|
2019-08-29 13:04:01 +03:00
|
|
|
input_data = input_path.open("r", encoding="utf-8").read()
|
2018-11-30 22:16:14 +03:00
|
|
|
if converter == "auto":
|
2017-10-10 04:06:28 +03:00
|
|
|
converter = input_path.suffix[1:]
|
2019-08-29 13:04:01 +03:00
|
|
|
if converter == "ner" or converter == "iob":
|
|
|
|
converter_autodetect = autodetect_ner_format(input_data)
|
|
|
|
if converter_autodetect == "ner":
|
|
|
|
msg.info("Auto-detected token-per-line NER format")
|
|
|
|
converter = converter_autodetect
|
|
|
|
elif converter_autodetect == "iob":
|
|
|
|
msg.info("Auto-detected sentence-per-line NER format")
|
|
|
|
converter = converter_autodetect
|
|
|
|
else:
|
2019-08-31 14:39:06 +03:00
|
|
|
msg.warn(
|
2020-06-21 22:35:01 +03:00
|
|
|
"Can't automatically detect NER format. Conversion may not "
|
|
|
|
"succeed. See https://spacy.io/api/cli#convert"
|
2019-08-31 14:39:06 +03:00
|
|
|
)
|
2017-10-27 15:38:39 +03:00
|
|
|
if converter not in CONVERTERS:
|
2019-12-22 03:53:56 +03:00
|
|
|
msg.fail(f"Can't find converter for {converter}", exits=1)
|
2019-12-11 20:20:49 +03:00
|
|
|
ner_map = None
|
|
|
|
if ner_map_path is not None:
|
|
|
|
ner_map = srsly.read_json(ner_map_path)
|
2018-11-30 22:16:14 +03:00
|
|
|
# Use converter function to convert data
|
2017-10-10 04:06:28 +03:00
|
|
|
func = CONVERTERS[converter]
|
2019-08-31 14:39:06 +03:00
|
|
|
data = func(
|
|
|
|
input_data,
|
|
|
|
n_sents=n_sents,
|
|
|
|
seg_sents=seg_sents,
|
2020-01-29 19:44:25 +03:00
|
|
|
append_morphology=morphology,
|
|
|
|
merge_subtokens=merge_subtokens,
|
2019-08-31 14:39:06 +03:00
|
|
|
lang=lang,
|
|
|
|
model=model,
|
2020-06-21 22:35:01 +03:00
|
|
|
no_print=silent,
|
2019-12-11 20:20:49 +03:00
|
|
|
ner_map=ner_map,
|
2019-08-31 14:39:06 +03:00
|
|
|
)
|
2018-11-30 22:16:14 +03:00
|
|
|
if output_dir != "-":
|
|
|
|
# Export data to a file
|
2019-12-22 03:53:56 +03:00
|
|
|
suffix = f".{file_type}"
|
2018-11-30 22:16:14 +03:00
|
|
|
output_file = Path(output_dir) / Path(input_path.parts[-1]).with_suffix(suffix)
|
|
|
|
if file_type == "json":
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 03:28:22 +03:00
|
|
|
srsly.write_json(output_file, data)
|
2018-11-30 22:16:14 +03:00
|
|
|
elif file_type == "jsonl":
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 03:28:22 +03:00
|
|
|
srsly.write_jsonl(output_file, data)
|
2019-03-09 01:15:23 +03:00
|
|
|
elif file_type == "msg":
|
|
|
|
srsly.write_msgpack(output_file, data)
|
2019-12-22 03:53:56 +03:00
|
|
|
msg.good(f"Generated output file ({len(data)} documents): {output_file}")
|
2018-11-30 22:16:14 +03:00
|
|
|
else:
|
|
|
|
# Print to stdout
|
|
|
|
if file_type == "json":
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 03:28:22 +03:00
|
|
|
srsly.write_json("-", data)
|
2018-11-30 22:16:14 +03:00
|
|
|
elif file_type == "jsonl":
|
💫 Replace ujson, msgpack and dill/pickle/cloudpickle with srsly (#3003)
Remove hacks and wrappers, keep code in sync across our libraries and move spaCy a few steps closer to only depending on packages with binary wheels 🎉
See here: https://github.com/explosion/srsly
Serialization is hard, especially across Python versions and multiple platforms. After dealing with many subtle bugs over the years (encodings, locales, large files) our libraries like spaCy and Prodigy have steadily grown a number of utility functions to wrap the multiple serialization formats we need to support (especially json, msgpack and pickle). These wrapping functions ended up duplicated across our codebases, so we wanted to put them in one place.
At the same time, we noticed that having a lot of small dependencies was making maintainence harder, and making installation slower. To solve this, we've made srsly standalone, by including the component packages directly within it. This way we can provide all the serialization utilities we need in a single binary wheel.
srsly currently includes forks of the following packages:
ujson
msgpack
msgpack-numpy
cloudpickle
* WIP: replace json/ujson with srsly
* Replace ujson in examples
Use regular json instead of srsly to make code easier to read and follow
* Update requirements
* Fix imports
* Fix typos
* Replace msgpack with srsly
* Fix warning
2018-12-03 03:28:22 +03:00
|
|
|
srsly.write_jsonl("-", data)
|
2019-08-29 13:04:01 +03:00
|
|
|
|
|
|
|
|
2020-06-21 22:35:01 +03:00
|
|
|
def autodetect_ner_format(input_data: str) -> str:
|
2019-08-29 13:04:01 +03:00
|
|
|
# guess format from the first 20 lines
|
|
|
|
lines = input_data.split("\n")[:20]
|
|
|
|
format_guesses = {"ner": 0, "iob": 0}
|
|
|
|
iob_re = re.compile(r"\S+\|(O|[IB]-\S+)")
|
|
|
|
ner_re = re.compile(r"\S+\s+(O|[IB]-\S+)$")
|
|
|
|
for line in lines:
|
|
|
|
line = line.strip()
|
|
|
|
if iob_re.search(line):
|
|
|
|
format_guesses["iob"] += 1
|
|
|
|
if ner_re.search(line):
|
|
|
|
format_guesses["ner"] += 1
|
|
|
|
if format_guesses["iob"] == 0 and format_guesses["ner"] > 0:
|
|
|
|
return "ner"
|
|
|
|
if format_guesses["ner"] == 0 and format_guesses["iob"] > 0:
|
|
|
|
return "iob"
|
|
|
|
return None
|