diff --git a/website/docs/usage/embeddings-transformers.mdx b/website/docs/usage/embeddings-transformers.mdx index cf80822fb..068db29f2 100644 --- a/website/docs/usage/embeddings-transformers.mdx +++ b/website/docs/usage/embeddings-transformers.mdx @@ -745,14 +745,17 @@ To benefit from pretraining, your training step needs to know to initialize its this by setting `initialize.init_tok2vec` to the filename of the `.bin` file that you want to use from pretraining. -A pretraining step that runs for 5 epochs with an output path of `pretrain/`, as -an example, produces `pretrain/model0.bin` through `pretrain/model4.bin`. To -make use of the final output, you could fill in this value in your config file: +Similar to training, pretraining produces a `model-last.bin` file which is the +last iteration of the trained weights which you can use to initialize your +`tok2vec` layer. Additionally, you can configure `n_save_epoch` to tell +pretraining in which epoch interval it should save the current training +progress. To make use of the final output, you could fill in this value in your +config file: ```ini {title="config.cfg"} [paths] -init_tok2vec = "pretrain/model4.bin" +init_tok2vec = "pretrain/model-last.bin" [initialize] init_tok2vec = ${paths.init_tok2vec}