TY - GEN
T1 - Minipatch learning as implicit ridge-like regularization
AU - Yao, Tianyi
AU - Lejeune, Daniel
AU - Javadi, Hamid
AU - Baraniuk, Richard G.
AU - Allen, Genevera I.
N1 - Funding Information:
ACKNOWLEDGMENT TY and GIA acknowledge support from NSF DMS-1554821, NSF NeuroNex-1707400, and NIH 1R01GM140468. DL, HJ, and RB were supported by NSF CCF-1911094, IIS-1838177, and IIS1730574; ONR N00014-18-12571 and N00014-17-1-2551; AFOSR FA9550-18-1-0478; DARPA G001534-7500; and a Vannevar Bush Faculty Fellowship, ONR N00014-18-1-2047.
Publisher Copyright:
© 2021 IEEE.
PY - 2021/1
Y1 - 2021/1
N2 - Ridge-like regularization often leads to improved generalization performance of machine learning models by mitigating overfitting. While ridge-regularized machine learning methods are widely used in many important applications, direct training via optimization could become challenging in huge data scenarios with millions of examples and features. We tackle such challenges by proposing a general approach that achieves ridge-like regularization through implicit techniques named Minipatch Ridge (MPRidge). Our approach is based on taking an ensemble of coefficients of unregularized learners trained on many tiny, random subsamples of both the examples and features of the training data, which we call minipatches. We empirically demonstrate that MPRidge induces an implicit ridge-like regularizing effect and performs nearly the same as explicit ridge regularization for a general class of predictors including logistic regression, SVM, and robust regression. Embarrassingly parallelizable, MPRidge provides a computationally appealing alternative to inducing ridge-like regularization for improving generalization performance in challenging big-data settings.
AB - Ridge-like regularization often leads to improved generalization performance of machine learning models by mitigating overfitting. While ridge-regularized machine learning methods are widely used in many important applications, direct training via optimization could become challenging in huge data scenarios with millions of examples and features. We tackle such challenges by proposing a general approach that achieves ridge-like regularization through implicit techniques named Minipatch Ridge (MPRidge). Our approach is based on taking an ensemble of coefficients of unregularized learners trained on many tiny, random subsamples of both the examples and features of the training data, which we call minipatches. We empirically demonstrate that MPRidge induces an implicit ridge-like regularizing effect and performs nearly the same as explicit ridge regularization for a general class of predictors including logistic regression, SVM, and robust regression. Embarrassingly parallelizable, MPRidge provides a computationally appealing alternative to inducing ridge-like regularization for improving generalization performance in challenging big-data settings.
KW - Ensemble learning
KW - Implicit regularization
KW - Ridge-like regularization
UR - http://www.scopus.com/inward/record.url?scp=85102976524&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85102976524&partnerID=8YFLogxK
U2 - 10.1109/BigComp51126.2021.00021
DO - 10.1109/BigComp51126.2021.00021
M3 - Conference contribution
AN - SCOPUS:85102976524
T3 - Proceedings - 2021 IEEE International Conference on Big Data and Smart Computing, BigComp 2021
SP - 65
EP - 68
BT - Proceedings - 2021 IEEE International Conference on Big Data and Smart Computing, BigComp 2021
A2 - Unger, Herwig
A2 - Kim, Jinho
A2 - Kang, U
A2 - So-In, Chakchai
A2 - Du, Junping
A2 - Saad, Walid
A2 - Ha, Young-guk
A2 - Wagner, Christian
A2 - Bourgeois, Julien
A2 - Sathitwiriyawong, Chanboon
A2 - Kwon, Hyuk-Yoon
A2 - Leung, Carson
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 IEEE International Conference on Big Data and Smart Computing, BigComp 2021
Y2 - 17 January 2021 through 20 January 2021
ER -