Maximum Relative Margin and Data-Dependent Regularization

Journal of Machine Learning Research(2010)

引用 89|浏览19
暂无评分
摘要
Leading classification methods such as support vector machines (SVMs) and their counterparts achieve strong generalization performance by maximizing the margin of separation between data classes. While the maximum margin approach has achieved promising performance, this article identifies its sensitivity to affine transformations of the data and to directions with large data spread. Maximum margin solutions may be misled by the spread of data and preferentially separate classes along large spread directions. This article corrects these weaknesses by measuring margin not in the absolute sense but rather only relative to the spread of data in any projection direction. Maximum relative margin corresponds to a data-dependent regularization on the classification function while maximum absolute margin corresponds to an l2 norm constraint on the classification function. Interestingly, the proposed improvements only require simple extensions to existing maximum margin formulations and preserve the computational efficiency of SVMs. Through the maximization of relative margin, surprising performance gains are achieved on real-world problems such as digit, text classification and on several other benchmark data sets. In addition, risk bounds are derived for the new formulation based on Rademacher averages.
更多
查看译文
关键词
data class,maximum relative margin,kernel methods,mnist.,benchmark data set,classification function,large margin,relative margin,maximum margin solution,support vector machines,maximum absolute margin corresponds,data-dependent regularization,maximum margin approach,large data spread,maximum margin formulation,rademacher complexity,maximum relative margin corresponds,kernel method,affine transformation,support vector machine
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要