Knowledge Distillation 1 Paper Review. Training data-efficient image transformers and distillation through attention@ICML’ 2021 Aug 18, 2021