A Comparative Study of Parameter Estimation Methods for Statistical Natural Language Processing

ACL(2007)

引用 71|浏览149
暂无评分
摘要
This paper presents a comparative study of five parameter estimation algorithms on four NLP tasks. Three of the five algorithms are well-known in the computational linguistics community: Maximum Entropy (ME) estima- tion with L2 regularization, the Averaged Perceptron (AP), and Boosting. We also in- vestigate ME estimation with L1 regularization using a novel optimization algorithm, and BLasso, which is a version of Boosting with Lasso (L1) regularization. We first investigate all of our estimators on two re-ranking tasks: a parse selection task and a language model (LM) adaptation task. Then we apply the best of these estimators to two additional tasks involving conditional sequence models: a Conditional Markov Model (CMM) for part of speech tagging and a Conditional Random Field (CRF) for Chinese word segmentation. Our experiments show that across tasks, three of the estimators — ME estimation with L1 or L2 regularization, and AP — are in a near sta- tistical tie for first place.
更多
查看译文
关键词
conditional random field,markov model,parameter estimation,maximum entropy,language model
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要