Sparse Coding And Autoencoders

2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT)(2018)

引用 16|浏览65
暂无评分
摘要
In this work we study the landscape of squared loss of an Autoencoder when the data generative model is that of "Sparse Coding"/" Dictionary Learning". The neural net considered is an R-n -> R-n mapping and has a single ReLU activation layer of size h > n. The net has access to vectors y is an element of R-n obtained as y - A*x* where x* is an element of R-h are sparse high dimensional vectors and A* is an element of R-n x h is an overcomplete incoherent matrix. Under very mild distributional assumptions on x*, we prove that the norm of the expected gradient of the squared loss function is asymptotically (in sparse code dimension) negligible for all points in a small neighborhood of A*. This is supported with experimental evidence using synthetic data. We conduct experiments to suggest that A* sits at the bottom of a well in the landscape and we also give experiments showing that gradient descent on this loss function gets columnwise very close to the original dictionary even with far enough initialization. Along the way we prove that a layer of ReLU gates can be set up to automatically recover the support of the sparse codes. Since this property holds independent of the loss function we believe that it could be of independent interest.
更多
查看译文
关键词
sparse high dimensional vectors,overcomplete incoherent matrix,squared loss function,sparse code dimension,synthetic data,gradient descent,ReLU gates,sparse codes,data generative model,neural net,single ReLU activation layer,dictionary learning,autoencoders,sparse coding
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要