Abbreviated text input using language modeling

Natural Language Engineering(2007)

引用 44|浏览7
暂无评分
摘要
We address the problem of improving the eciency of natural language text input un- der degraded conditions (for instance, on mobile computing devices or by disabled users), by taking advantage of the informational redundancy in natural language. Previous ap- proaches to this problem have been based on the idea of prediction of the text, but these require the user to take overt action to verify or select the system's predictions. We pro- pose taking advantage of the duality between prediction and compression. We allow the user to enter text in compressed form, in particular, using a simple stipulated abbrevia- tion method that reduces characters by 26.4%, yet is simple enough that it can be learned easily and generated relatively fluently. We decode the abbreviated text using a statisti- cal generative model of abbreviation, with a residual word error rate of 3.3%. The chief component of this model is an n-gram language model. Because the system's operation is completely independent from the user's, the overhead from cognitive task switching and attending to the system's actions online is eliminated, opening up the possibility that the compression-based method can achieve text input eciency improvements where the prediction-based methods have not. We report the results of a user study evaluating this method.
更多
查看译文
关键词
computer science,general relativity,word error rate,language modeling,task switching,language model,mobile computer,natural language
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要