NetDistiller: Empowering Tiny Deep Learning via In Situ Distillation

IEEE MICRO(2023)

引用 0|浏览9
暂无评分
摘要
Boosting the task accuracy of tiny neural networks (TNNs) has become a fundamental challenge for enabling the deployment of TNNs on edge devices, which are constrained by strict limitations in terms of memory, computation, bandwidth, and power supply. To this end, we propose a framework called NetDistiller to boost the achievable accuracy of TNNs by treating them as subnetworks of a weight-sharing teacher constructed by expanding the number of channels of the TNN. Specifically, the target TNN model is jointly trained with the weight-sharing teacher model via 1) gradient surgery to tackle the gradient conflicts between them and 2) uncertainty-aware distillation to mitigate the overfitting of the teacher model. Extensive experiments across diverse tasks validate NetDistiller's effectiveness in boosting TNNs' achievable accuracy over state-of-the-art methods. Our code is available at https://github.com/GATECH-EIC/NetDistiller.
更多
查看译文
关键词
tiny deep learning,deep learning,netdistiller,in-situ
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要