TY - JOUR
T1 - Smoothly Giving up
T2 - 26th International Conference on Artificial Intelligence and Statistics, AISTATS 2023
AU - Sypherd, Tyler
AU - Stromberg, Nathan
AU - Nock, Richard
AU - Berisha, Visar
AU - Sankar, Lalitha
N1 - Funding Information: We thank the anonymous reviewers for their comments, and Monica Welfert at Arizona State University for her contributions to the preliminary code. This work is supported in part by NSF grants SCH-2205080, CIF-1901243, CIF-2134256, CIF-2007688, CIF-1815361, a Google AI for Social Good grant, and an Office of Naval Research grant N00014-21-1-2615. This research is based on survey results from Carnegie Mellon University's Delphi Group. Funding Information: We thank the anonymous reviewers for their comments, and Monica Welfert at Arizona State University for her contributions to the preliminary code. This work is supported in part by NSF grants SCH-2205080, CIF-1901243, CIF-2134256, CIF-2007688, CIF-1815361, a Google AI for Social Good grant, and an Office of Naval Research grant N00014-21-1-2615. This research is based on survey results from Carnegie Mellon University’s Delphi Group. Publisher Copyright: Copyright © 2023 by the author(s)
PY - 2023
Y1 - 2023
N2 - There is a growing need for models that are interpretable and have reduced energy/computational cost (e.g., in health care analytics and federated learning). Examples of algorithms to train such models include logistic regression and boosting. However, one challenge facing these algorithms is that they provably suffer from label noise; this has been attributed to the joint interaction between oft-used convex loss functions and simpler hypothesis classes, resulting in too much emphasis being placed on outliers. In this work, we use the margin-based α-loss, which continuously tunes between canonical convex and quasi-convex losses, to robustly train simple models. We show that the α hyperparameter smoothly introduces non-convexity and offers the benefit of “giving up” on noisy training examples. We also provide results on the Long-Servedio dataset for boosting and a COVID-19 survey dataset for logistic regression, highlighting the efficacy of our approach across multiple relevant domains.
AB - There is a growing need for models that are interpretable and have reduced energy/computational cost (e.g., in health care analytics and federated learning). Examples of algorithms to train such models include logistic regression and boosting. However, one challenge facing these algorithms is that they provably suffer from label noise; this has been attributed to the joint interaction between oft-used convex loss functions and simpler hypothesis classes, resulting in too much emphasis being placed on outliers. In this work, we use the margin-based α-loss, which continuously tunes between canonical convex and quasi-convex losses, to robustly train simple models. We show that the α hyperparameter smoothly introduces non-convexity and offers the benefit of “giving up” on noisy training examples. We also provide results on the Long-Servedio dataset for boosting and a COVID-19 survey dataset for logistic regression, highlighting the efficacy of our approach across multiple relevant domains.
UR - http://www.scopus.com/inward/record.url?scp=85165195313&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85165195313&partnerID=8YFLogxK
M3 - Conference article
SN - 2640-3498
VL - 206
SP - 5376
EP - 5410
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
Y2 - 25 April 2023 through 27 April 2023
ER -