213. Knowledge Distillation
A technique where a smaller model is trained to replicate the behavior of a larger, more complex model.
Last updated
A technique where a smaller model is trained to replicate the behavior of a larger, more complex model.
Last updated