Online Learning of a Memory for Learning Rates

2018 IEEE International Conference on Robotics and Automation (ICRA)(2018)

引用 26|浏览159
暂无评分
摘要
The promise of learning to learn for robotics rests on the hope that by extracting some information about the learning process itself we can speed up subsequent similar learning tasks. Here, we introduce a computationally efficient online meta-learning algorithm that builds and optimizes a memory model of the optimal learning rate landscape from previously observed gradient behaviors. While performing task specific optimization, this memory of learning rates predicts how to scale currently observed gradients. After applying the gradient scaling our meta-learner updates its internal memory based on the observed effect its prediction had. Our meta-learner can be combined with any gradient-based optimizer, learns on the fly and can be transferred to new optimization tasks. In our evaluations we show that our meta-learning algorithm speeds up learning of MNIST classification and a variety of learning control tasks, either in batch or online learning settings.
更多
查看译文
关键词
learning rates,learning process,memory model,optimal learning rate landscape,task specific optimization,meta-learner,internal memory,optimization tasks,meta-learning algorithm speeds,learning control tasks,online learning settings,gradient behaviors,gradient-based optimizer,MNIST classification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要