Attention-based Feature Interaction for Efficient Online Knowledge Distillation

2021 IEEE International Conference on Data Mining (ICDM)(2021)

引用 3|浏览21
暂无评分
摘要
Existing online knowledge distillation (KD) methods solve the dependency problem of the high-capacity teacher model via mutual learning and ensemble learning. But they focus on the utilization of logits information in the last few layers and fail to construct a strong teacher model to better supervise student networks, leading to the inefficiency of KD. In this work, we propose a simple but effect...
更多
查看译文
关键词
Representation learning,Knowledge engineering,Learning systems,Fuses,Conferences,Education,Focusing
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要