The International Journal of Biostatistics
Ed. by Chambaz, Antoine / Hubbard, Alan E. / van der Laan, Mark J.
2 Issues per year
IMPACT FACTOR 2013: 0.948
SCImago Journal Rank (SJR): 1.039
Source Normalized Impact per Paper (SNIP): 0.746
Mathematical Citation Quotient 2013: 0.04
Volume 7 (2011)
Volume 5 (2009)
Volume 4 (2008)
Volume 3 (2007)
Volume 2 (2006)
Volume 1 (2005)
Most Downloaded Articles
- An Introduction to Causal Inference by Pearl, Judea
- Sample Size Estimation for Repeated Measures Analysis in Randomized Clinical Trials with Missing Data by Lu, Kaifeng/ Luo, Xiaohui and Chen, Pei-Yun
- Survival Models in Health Economic Evaluations: Balancing Fit and Parsimony to Improve Prediction by Jackson, Christopher H/ Sharples, Linda D and Thompson, Simon G
- Accuracy of Conventional and Marginal Structural Cox Model Estimators: A Simulation Study by Xiao, Yongling/ Abrahamowicz, Michal and Moodie, Erica E. M.
- Model Checking with Residuals for g-estimation of Optimal Dynamic Treatment Regimes by Rich, Benjamin/ Moodie, Erica E. M./ Stephens, David A and Platt, Robert W
HingeBoost: ROC-Based Boost for Classification and Variable Selection
1Connecticut Children’s Medical Center and University of Connecticut School of Medicine
Citation Information: The International Journal of Biostatistics. Volume 7, Issue 1, Pages 1–30, ISSN (Online) 1557-4679, DOI: 10.2202/1557-4679.1304, February 2011
- Published Online:
In disease classification, a traditional technique is the receiver operative characteristic (ROC) curve and the area under the curve (AUC). With high-dimensional data, the ROC techniques are needed to conduct classification and variable selection. The current ROC methods do not explicitly incorporate unequal misclassification costs or do not have a theoretical grounding for optimizing the AUC. Empirical studies in the literature have demonstrated that optimizing the hinge loss can maximize the AUC approximately. In theory, minimizing the hinge rank loss is equivalent to minimizing the AUC in the asymptotic limit. In this article, we propose a novel nonparametric method HingeBoost to optimize a weighted hinge loss incorporating misclassification costs. HingeBoost can be used to construct linear and nonlinear classifiers. The estimation and variable selection for the hinge loss are addressed by a new boosting algorithm. Furthermore, the proposed twin HingeBoost can select more sparse predictors. Some properties of HingeBoost are studied as well. To compare HingeBoost with existing classification methods, we present empirical study results using data from simulations and a prostate cancer study with mass spectrometry-based proteomics.