-
The Experience
-
About Stanford GSB
About Our Degree Programs
-
-
The Programs
-
Full-Time Degree Programs
Non-Degree & Certificate Programs
-
-
Faculty & Research
-
Faculty
Faculty Research
Research Hub
Centers & Institutes
-
-
Insights
-
Topics
-
-
Alumni
-
Welcome, Alumni
-
-
Events
-
Admission Events & Information Sessions
-
Stable Prediction with Model Misspecification and Agnostic Distribution Shift
Stable Prediction with Model Misspecification and Agnostic Distribution Shift
January 2020Working Paper No. 3898
For many machine learning algorithms, two main assumptions are required to guarantee performance. One is that the test data are drawn from the same distribution as the training data, and the other is that the model is correctly specified. In real applications, however, we often have little prior knowledge on the test data and on the underlying true model. Under model misspecification, agnostic distribution shift between training and test data leads to inaccuracy of parameter estimation and instability of prediction across unknown test data. To address these problems, we propose a novel Decorrelated Weighting Regression algorithm which jointly optimizes a variable decorrelation regularizer and a weighted regression model. The variable decorrelation regularizer estimates a weight for each sample such that variables are decorrelated on the weighted training data. Then, these weights are used in the weighted regression to improve the accuracy of estimation on the effect of each variable, thus help to improve the stability of prediction across unknown test data. Extensive experiments clearly demonstrate that our DWR algorithm can significantly improve the accuracy of parameter estimation and stability of prediction with model misspecification and agnostic distribution shift.