Technique
Knowledge distillation
Training a smaller "student" model to match the outputs of a larger "teacher" model, producing a cheaper model that retains much of the teacher's quality.
Technique
Training a smaller "student" model to match the outputs of a larger "teacher" model, producing a cheaper model that retains much of the teacher's quality.
We use cookies
Anonymous analytics help us improve the site. You can opt out anytime. Learn more