mirror of
https://github.com/explosion/spaCy.git
synced 2025-07-10 08:12:24 +03:00
Load vectors in 'spacy train'
This commit is contained in:
parent
7852a68a75
commit
2aff3c4b5a
|
@ -80,16 +80,20 @@ def train(
|
||||||
msg.info("Using CPU")
|
msg.info("Using CPU")
|
||||||
msg.info(f"Loading config and nlp from: {config_path}")
|
msg.info(f"Loading config and nlp from: {config_path}")
|
||||||
config = Config().from_disk(config_path)
|
config = Config().from_disk(config_path)
|
||||||
|
if config.get("training", {}).get("seed") is not None:
|
||||||
|
fix_random_seed(config["training"]["seed"])
|
||||||
with show_validation_error():
|
with show_validation_error():
|
||||||
nlp, config = util.load_model_from_config(config, overrides=config_overrides)
|
nlp, config = util.load_model_from_config(config, overrides=config_overrides)
|
||||||
if config["training"]["base_model"]:
|
if config["training"]["base_model"]:
|
||||||
base_nlp = util.load_model(config["training"]["base_model"])
|
|
||||||
# TODO: do something to check base_nlp against regular nlp described in config?
|
# TODO: do something to check base_nlp against regular nlp described in config?
|
||||||
nlp = base_nlp
|
# If everything matches it will look something like:
|
||||||
|
# base_nlp = util.load_model(config["training"]["base_model"])
|
||||||
|
# nlp = base_nlp
|
||||||
|
raise NotImplementedError("base_model not supported yet.")
|
||||||
|
if config["training"]["vectors"] is not None:
|
||||||
|
util.load_vectors_into_model(nlp, config["training"]["vectors"])
|
||||||
verify_config(nlp)
|
verify_config(nlp)
|
||||||
raw_text, tag_map, morph_rules, weights_data = load_from_paths(config)
|
raw_text, tag_map, morph_rules, weights_data = load_from_paths(config)
|
||||||
if config["training"]["seed"] is not None:
|
|
||||||
fix_random_seed(config["training"]["seed"])
|
|
||||||
if config["training"]["use_pytorch_for_gpu_memory"]:
|
if config["training"]["use_pytorch_for_gpu_memory"]:
|
||||||
# It feels kind of weird to not have a default for this.
|
# It feels kind of weird to not have a default for this.
|
||||||
use_pytorch_for_gpu_memory()
|
use_pytorch_for_gpu_memory()
|
||||||
|
|
Loading…
Reference in New Issue
Block a user