213. Knowledge Distillation

A technique where a smaller model is trained to replicate the behavior of a larger, more complex model.

Last updated