What is the purpose of knowledge distillation in model compression?

Machine Learning Hard

Machine Learning — Hard

What is the purpose of knowledge distillation in model compression?

Key points

  • Knowledge distillation transfers knowledge through soft probabilities
  • The goal is to compress a model while maintaining performance
  • Student model mimics the teacher model's behavior
  • Soft labels are used instead of hard labels
  • Efficient way to compress models without significant loss in performance

Ready to go further?

Related questions