mirror of
https://github.com/explosion/spaCy.git
synced 2025-03-03 10:55:52 +03:00
Fix learn rate for non-transformer
This commit is contained in:
parent
465785a672
commit
4b7abaafdb
|
@ -186,11 +186,14 @@ accumulate_gradient = {{ transformer["size_factor"] }}
|
|||
[training.optimizer]
|
||||
@optimizers = "Adam.v1"
|
||||
|
||||
|
||||
{% if use_transformer -%}
|
||||
[training.optimizer.learn_rate]
|
||||
@schedules = "warmup_linear.v1"
|
||||
warmup_steps = 250
|
||||
total_steps = 20000
|
||||
initial_rate = 5e-5
|
||||
{% endif %}
|
||||
|
||||
[training.train_corpus]
|
||||
@readers = "spacy.Corpus.v1"
|
||||
|
|
Loading…
Reference in New Issue
Block a user