SMART GRADIENT - AN ADAPTIVE TECHNIQUE FOR IMPROVING GRADIENT ESTIMATION

FOUNDATIONS OF DATA SCIENCE(2022)

引用 5|浏览2
暂无评分
摘要
Computing the gradient of a function provides fundamental information about its behavior. This information is essential for several applications and algorithms across various fields. One common application that requires gradients are optimization techniques such as stochastic gradient descent, Newton's method and trust region methods. However, these methods usually require a numerical computation of the gradient at every iteration of the method which is prone to numerical errors. We propose a simple limited-memory technique for improving the accuracy of a numerically computed gradient in this gradient-based optimization framework by exploiting (1) a coordinate transformation of the gradient and (2) the history of previously taken descent directions. The method is verified empirically by extensive experimentation on both test functions and on real data applications. The proposed method is implemented in the R package smartGrad and in C++.
更多
查看译文
关键词
Adaptive technique, gradient estimation, numerical gradient, optimization, vanilla gradient descent
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要