A study on effects of implicit and explicit language model information for DBLSTM-CTC based handwriting recognition

International Conference on Document Analysis and Recognition(2015)

引用 33|浏览88
暂无评分
摘要
Deep Bidirectional Long Short-Term Memory (DBLSTM) with a Connectionist Temporal Classification (CTC) output layer has been established as one of the state-of-the-art solutions for handwriting recognition. It is well-known that the DBLSTM trained by using a CTC objective function will learn both local character image dependency for character modeling and long-range contextual dependency for implicit language modeling. In this paper, we study the effects of implicit and explicit language model information for DBLSTM-CTC based handwriting recognition by comparing the performance of using or without using an explicit language model in decoding. It is observed that even using one million lines of training sentences to train the DBLSTM, using an explicit language model is still helpful. To deal with such a large-scale training problem, a GPU-based training tool has been developed for CTC training of DBLSTM by using a mini-batch based epochwise Back Propagation Through Time (BPTT) algorithm.
更多
查看译文
关键词
DBLSTM-CTC based handwriting recognition,implicit language model information,explicit language model information,deep bidirectional long short-term memory,connectionist temporal classification output layer,DBLSTM training,CTC objective function,local character image dependency,character modeling,long-range contextual dependency,training sentences,large-scale training problem,CTC training,GPU-based training tool,mini-batch based epoch wise back propagation through time algorithm
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要