mirror of
https://github.com/explosion/spaCy.git
synced 2025-10-25 13:11:03 +03:00
* Add `TrainablePipe.{distill,get_teacher_student_loss}`
This change adds two methods:
- `TrainablePipe::distill` which performs a training step of a
student pipe on a teacher pipe, giving a batch of `Doc`s.
- `TrainablePipe::get_teacher_student_loss` computes the loss
of a student relative to the teacher.
The `distill` or `get_teacher_student_loss` methods are also implemented
in the tagger, edit tree lemmatizer, and parser pipes, to enable
distillation in those pipes and as an example for other pipes.
* Fix stray `Beam` import
* Fix incorrect import
* Apply suggestions from code review
Co-authored-by: Sofie Van Landeghem <svlandeg@users.noreply.github.com>
* Apply suggestions from code review
Co-authored-by: Sofie Van Landeghem <svlandeg@users.noreply.github.com>
* TrainablePipe.distill: use `Iterable[Example]`
* Add Pipe.is_distillable method
* Add `validate_distillation_examples`
This first calls `validate_examples` and then checks that the
student/teacher tokens are the same.
* Update distill documentation
* Add distill documentation for all pipes that support distillation
* Fix incorrect identifier
* Apply suggestions from code review
Co-authored-by: Sofie Van Landeghem <svlandeg@users.noreply.github.com>
* Add comment to explain `is_distillable`
Co-authored-by: Sofie Van Landeghem <svlandeg@users.noreply.github.com>
|
||
|---|---|---|
| .. | ||
| 101 | ||
| _benchmarks-models.mdx | ||
| embeddings-transformers.mdx | ||
| facts-figures.mdx | ||
| index.mdx | ||
| layers-architectures.mdx | ||
| linguistic-features.mdx | ||
| models.mdx | ||
| processing-pipelines.mdx | ||
| projects.mdx | ||
| rule-based-matching.mdx | ||
| saving-loading.mdx | ||
| spacy-101.mdx | ||
| training.mdx | ||
| v2-1.mdx | ||
| v2-2.mdx | ||
| v2-3.mdx | ||
| v2.mdx | ||
| v3-1.mdx | ||
| v3-2.mdx | ||
| v3-3.mdx | ||
| v3-4.mdx | ||
| v3.mdx | ||
| visualizers.mdx | ||