Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices
NIPS 2020(2020)
摘要
We present an extension of the conditional gradient method to problems whose feasible sets are convex cones. We provide a convergence analysis for the method and for variants with nonconvex objectives, and we extend the analysis to practical cases with effective line search strategies. For the specific case of the positive semidefinite cone, we present a memory-efficient version based on randomized matrix sketches and advocate a heuristic greedy step that greatly improves its practical performance. Numerical results on phase retrieval and matrix completion problems indicate that our method can offer substantial advantages over traditional conditional gradient and Burer-Monteiro approaches.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络