mirror of
https://github.com/explosion/spaCy.git
synced 2025-03-03 19:08:06 +03:00
Fix learn rate for non-transformer
This commit is contained in:
parent
465785a672
commit
4b7abaafdb
|
@ -186,11 +186,14 @@ accumulate_gradient = {{ transformer["size_factor"] }}
|
||||||
[training.optimizer]
|
[training.optimizer]
|
||||||
@optimizers = "Adam.v1"
|
@optimizers = "Adam.v1"
|
||||||
|
|
||||||
|
|
||||||
|
{% if use_transformer -%}
|
||||||
[training.optimizer.learn_rate]
|
[training.optimizer.learn_rate]
|
||||||
@schedules = "warmup_linear.v1"
|
@schedules = "warmup_linear.v1"
|
||||||
warmup_steps = 250
|
warmup_steps = 250
|
||||||
total_steps = 20000
|
total_steps = 20000
|
||||||
initial_rate = 5e-5
|
initial_rate = 5e-5
|
||||||
|
{% endif %}
|
||||||
|
|
||||||
[training.train_corpus]
|
[training.train_corpus]
|
||||||
@readers = "spacy.Corpus.v1"
|
@readers = "spacy.Corpus.v1"
|
||||||
|
|
Loading…
Reference in New Issue
Block a user