mirror of
https://github.com/explosion/spaCy.git
synced 2025-03-03 10:55:52 +03:00
Remove env_opt and simplfy default Optimizer
This commit is contained in:
parent
ab1d165bba
commit
37814b608d
|
@ -60,7 +60,6 @@ def evaluate(
|
||||||
fix_random_seed()
|
fix_random_seed()
|
||||||
if use_gpu >= 0:
|
if use_gpu >= 0:
|
||||||
require_gpu(use_gpu)
|
require_gpu(use_gpu)
|
||||||
util.set_env_log(False)
|
|
||||||
data_path = util.ensure_path(data_path)
|
data_path = util.ensure_path(data_path)
|
||||||
output_path = util.ensure_path(output)
|
output_path = util.ensure_path(output)
|
||||||
displacy_path = util.ensure_path(displacy_path)
|
displacy_path = util.ensure_path(displacy_path)
|
||||||
|
|
|
@ -48,7 +48,6 @@ def train_cli(
|
||||||
used to register custom functions and architectures that can then be
|
used to register custom functions and architectures that can then be
|
||||||
referenced in the config.
|
referenced in the config.
|
||||||
"""
|
"""
|
||||||
util.set_env_log(verbose)
|
|
||||||
verify_cli_args(config_path, output_path)
|
verify_cli_args(config_path, output_path)
|
||||||
overrides = parse_config_overrides(ctx.args)
|
overrides = parse_config_overrides(ctx.args)
|
||||||
import_code(code_path)
|
import_code(code_path)
|
||||||
|
|
|
@ -54,7 +54,6 @@ if TYPE_CHECKING:
|
||||||
from .vocab import Vocab # noqa: F401
|
from .vocab import Vocab # noqa: F401
|
||||||
|
|
||||||
|
|
||||||
_PRINT_ENV = False
|
|
||||||
OOV_RANK = numpy.iinfo(numpy.uint64).max
|
OOV_RANK = numpy.iinfo(numpy.uint64).max
|
||||||
LEXEME_NORM_LANGS = ["da", "de", "el", "en", "id", "lb", "pt", "ru", "sr", "ta", "th"]
|
LEXEME_NORM_LANGS = ["da", "de", "el", "en", "id", "lb", "pt", "ru", "sr", "ta", "th"]
|
||||||
|
|
||||||
|
@ -109,11 +108,6 @@ class SimpleFrozenDict(dict):
|
||||||
raise NotImplementedError(self.error)
|
raise NotImplementedError(self.error)
|
||||||
|
|
||||||
|
|
||||||
def set_env_log(value: bool) -> None:
|
|
||||||
global _PRINT_ENV
|
|
||||||
_PRINT_ENV = value
|
|
||||||
|
|
||||||
|
|
||||||
def lang_class_is_loaded(lang: str) -> bool:
|
def lang_class_is_loaded(lang: str) -> bool:
|
||||||
"""Check whether a Language class is already loaded. Language classes are
|
"""Check whether a Language class is already loaded. Language classes are
|
||||||
loaded lazily, to avoid expensive setup code associated with the language
|
loaded lazily, to avoid expensive setup code associated with the language
|
||||||
|
@ -602,27 +596,6 @@ def get_async(stream, numpy_array):
|
||||||
return array
|
return array
|
||||||
|
|
||||||
|
|
||||||
def env_opt(name: str, default: Optional[Any] = None) -> Optional[Any]:
|
|
||||||
if type(default) is float:
|
|
||||||
type_convert = float
|
|
||||||
else:
|
|
||||||
type_convert = int
|
|
||||||
if "SPACY_" + name.upper() in os.environ:
|
|
||||||
value = type_convert(os.environ["SPACY_" + name.upper()])
|
|
||||||
if _PRINT_ENV:
|
|
||||||
print(name, "=", repr(value), "via", "$SPACY_" + name.upper())
|
|
||||||
return value
|
|
||||||
elif name in os.environ:
|
|
||||||
value = type_convert(os.environ[name])
|
|
||||||
if _PRINT_ENV:
|
|
||||||
print(name, "=", repr(value), "via", "$" + name)
|
|
||||||
return value
|
|
||||||
else:
|
|
||||||
if _PRINT_ENV:
|
|
||||||
print(name, "=", repr(default), "by default")
|
|
||||||
return default
|
|
||||||
|
|
||||||
|
|
||||||
def read_regex(path: Union[str, Path]) -> Pattern:
|
def read_regex(path: Union[str, Path]) -> Pattern:
|
||||||
path = ensure_path(path)
|
path = ensure_path(path)
|
||||||
with path.open(encoding="utf8") as file_:
|
with path.open(encoding="utf8") as file_:
|
||||||
|
@ -1067,24 +1040,7 @@ class DummyTokenizer:
|
||||||
|
|
||||||
|
|
||||||
def create_default_optimizer() -> Optimizer:
|
def create_default_optimizer() -> Optimizer:
|
||||||
# TODO: Do we still want to allow env_opt?
|
return Adam()
|
||||||
learn_rate = env_opt("learn_rate", 0.001)
|
|
||||||
beta1 = env_opt("optimizer_B1", 0.9)
|
|
||||||
beta2 = env_opt("optimizer_B2", 0.999)
|
|
||||||
eps = env_opt("optimizer_eps", 1e-8)
|
|
||||||
L2 = env_opt("L2_penalty", 1e-6)
|
|
||||||
grad_clip = env_opt("grad_norm_clip", 10.0)
|
|
||||||
L2_is_weight_decay = env_opt("L2_is_weight_decay", False)
|
|
||||||
optimizer = Adam(
|
|
||||||
learn_rate,
|
|
||||||
L2=L2,
|
|
||||||
beta1=beta1,
|
|
||||||
beta2=beta2,
|
|
||||||
eps=eps,
|
|
||||||
grad_clip=grad_clip,
|
|
||||||
L2_is_weight_decay=L2_is_weight_decay,
|
|
||||||
)
|
|
||||||
return optimizer
|
|
||||||
|
|
||||||
|
|
||||||
def minibatch(items, size):
|
def minibatch(items, size):
|
||||||
|
|
Loading…
Reference in New Issue
Block a user