N2N Learning: Network to Network Compression via Policy Gradient Reinforcement Learning.

international conference on learning representations(2018)

引用 193|浏览26
暂无评分
摘要
While bigger and deeper neural network architectures continue to advance the state-of-the-art for many computer vision tasks, real-world adoption of these networks is impeded by hardware and speed constraints. Conventional model compression methods attempt to address this problem by modifying the architecture manually or using pre-defined heuristics. Since the space of all reduced architectures is very large, modifying the architecture of a deep neural network in this way is a difficult task. In this paper, we tackle this issue by introducing a principled method for learning reduced network architectures in a data-driven way using reinforcement learning. Our approach takes a larger u0027teacheru0027 network as input and outputs a compressed u0027studentu0027 network derived from the u0027teacheru0027 network. In the first stage of our method, a recurrent policy network aggressively removes layers from the large u0027teacheru0027 model. In the second stage, another recurrent policy network carefully reduces the size of each remaining layer. The resulting network is then evaluated to obtain a reward -- a score based on the accuracy and compression of the network. Our approach uses this reward signal with policy gradients to train the policies to find a locally optimal student network. Our experiments show that we can achieve compression rates of more than 10x for models such as ResNet-34 while maintaining similar performance to the input u0027teacheru0027 network. We also present a valuable transfer learning result which shows that policies which are pre-trained on smaller u0027teacheru0027 networks can be used to rapidly speed up training on larger u0027teacheru0027 networks.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要