Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Fast and Robust Online Inference with Stochastic Gradient Descent via Random Scaling

About

We develop a new method of online inference for a vector of parameters estimated by the Polyak-Ruppert averaging procedure of stochastic gradient descent (SGD) algorithms. We leverage insights from time series regression in econometrics and construct asymptotically pivotal statistics via random scaling. Our approach is fully operational with online data and is rigorously underpinned by a functional central limit theorem. Our proposed inference method has a couple of key advantages over the existing methods. First, the test statistic is computed in an online fashion with only SGD iterates and the critical values can be obtained without any resampling methods, thereby allowing for efficient implementation suitable for massive online data. Second, there is no need to estimate the asymptotic variance and our inference method is shown to be robust to changes in the tuning parameters for SGD algorithms in simulation experiments with synthetic data.

Sokbae Lee, Yuan Liao, Myung Hwan Seo, Youngki Shin• 2021

Related benchmarks

TaskDatasetResultRank
Linear regressionToeplitz Σ (n=10^5, d=200)
Covariance (%)97.02
15
Logistic RegressionToeplitz Σ (n=10^5, d=200)
Coverage76.38
15
Linear regressionToeplitz Σ n=10^5, d=5
Covariance92.83
15
Linear regressionToeplitz Σ (n=10^5, d=20)
Covariance (%)93.38
15
Logistic RegressionToeplitz Σ n=10^5, d=5
Coverage (%)92.67
15
Logistic RegressionToeplitz Σ (n=10^5, d=20)
Coverage (%)90.88
15
Linear regressionSynthetic d=5 (test)
Coverage99.86
6
Linear regressionSynthetic d=20
Coverage99.87
6
Logistic RegressionSynthetic d=5 (test)
Coverage99.76
6
Logistic RegressionSynthetic d=20
Coverage99.6
6
Showing 10 of 10 rows

Other info

Follow for update