HashNet: Deep Learning to Hash by Continuation

2017 IEEE International Conference on Computer Vision (ICCV)(2017)

引用 740|浏览315
暂无评分
摘要
Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and retrieval quality. Deep learning to hash, which improves retrieval quality by end-to-end representation learning and hash encoding, has received increasing attention recently. Subject to the ill-posed gradient difficulty in the optimization with sign activations, existing deep learning to hash methods need to first learn continuous representations and then generate binary hash codes in a separated binarization step, which suffer from substantial loss of retrieval quality. This work presents HashNet, a novel deep architecture for deep learning to hash by continuation method with convergence guarantees, which learns exactly binary hash codes from imbalanced similarity data. The key idea is to attack the ill-posed gradient problem in optimizing deep networks with non-smooth binary activations by continuation method, in which we begin from learning an easier network with smoothed activation function and let it evolve during the training, until it eventually goes back to being the original, difficult to optimize, deep network with the sign activation function. Comprehensive empirical evidence shows that HashNet can generate exactly binary hash codes and yield state-of-the-art multimedia retrieval performance on standard benchmarks.
更多
查看译文
关键词
deep network,HashNet,exactly binary hash codes,deep learning,approximate nearest neighbor search,retrieval quality,end-to-end representation learning,hash methods,continuous representations,deep architecture,continuation method,nonsmooth binary activations,multimedia retrieval,convergence guarantees,ill-posed gradient problem,smoothed activation function,multimedia retrieval performance,sign activation function
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要