Deep LearningArchitectures

Knowledge Distillation

Overview

A model compression technique where a smaller student model learns to mimic the behaviour of a larger teacher model.

More in Deep Learning