TY - JOUR
T1 - Hard or soft classification? large-margin unified machines
AU - Liu, Yufeng
AU - Zhang, Hao Helen
AU - Wu, Yichao
N1 - Funding Information:
Yufeng Liu is Associate Professor, Department of Statistics and Operations Research, Carolina Center for Genome Sciences, University of North Carolina, Chapel Hill, NC 27599 (E-mail: [email protected]). Hao Helen Zhang is Associate Professor and Yichao Wu is Assistant Professor, Department of Statistics, North Carolina State University, Raleigh, NC 27695-8203. The authors thank the editor, the associate editor, and three referees for their helpful suggestions that led to significant improvement of the article. The authors are partially supported by NSF grants DMS-0747575 (Liu), DMS-0645293 (Zhang) and DMS-0905561 (Wu), NIH grants NIH/NCI R01 CA-149569 (Liu and Wu), NIH/NCI P01 CA142538 (Liu and Zhang), and NIH/NCI R01 CA-085848 (Zhang).
PY - 2011/3
Y1 - 2011/3
N2 - Margin-based classifiers have been popular in both machine learning and statistics for classification problems. Among numerous classifiers, some are hard classifiers while some are soft ones. Soft classifiers explicitly estimate the class conditional probabilities and then perform classification based on estimated probabilities. In contrast, hard classifiers directly target the classification decision boundary without producing the probability estimation. These two types of classifiers are based on different philosophies and each has its own merits. In this article, we propose a novel family of large-margin classifiers, namely large-margin unified machines (LUMs), which covers a broad range of margin-based classifiers including both hard and soft ones. By offering a natural bridge from soft to hard classification, the LUM provides a unified algorithm to fit various classifiers and hence a convenient platform to compare hard and soft classification. Both theoretical consistency and numerical performance of LUMs are explored. Our numerical study sheds some light on the choice between hard and soft classifiers in various classification problems.
AB - Margin-based classifiers have been popular in both machine learning and statistics for classification problems. Among numerous classifiers, some are hard classifiers while some are soft ones. Soft classifiers explicitly estimate the class conditional probabilities and then perform classification based on estimated probabilities. In contrast, hard classifiers directly target the classification decision boundary without producing the probability estimation. These two types of classifiers are based on different philosophies and each has its own merits. In this article, we propose a novel family of large-margin classifiers, namely large-margin unified machines (LUMs), which covers a broad range of margin-based classifiers including both hard and soft ones. By offering a natural bridge from soft to hard classification, the LUM provides a unified algorithm to fit various classifiers and hence a convenient platform to compare hard and soft classification. Both theoretical consistency and numerical performance of LUMs are explored. Our numerical study sheds some light on the choice between hard and soft classifiers in various classification problems.
KW - Class probability estimation
KW - DWD
KW - Fisher consistency
KW - Regularization
KW - SVM
UR - http://www.scopus.com/inward/record.url?scp=79954529373&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=79954529373&partnerID=8YFLogxK
U2 - 10.1198/jasa.2011.tm10319
DO - 10.1198/jasa.2011.tm10319
M3 - Article
AN - SCOPUS:79954529373
SN - 0162-1459
VL - 106
SP - 166
EP - 177
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 493
ER -