Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Second-Order Stochastic Optimization for Machine Learning in Linear Time

About

First-order stochastic methods are the state-of-the-art in large-scale machine learning optimization owing to efficient per-iteration complexity. Second-order methods, while able to provide faster convergence, have been much less explored due to the high cost of computing the second-order information. In this paper we develop second-order stochastic methods for optimization problems in machine learning that match the per-iteration cost of gradient based methods, and in certain settings improve upon the overall running time over popular first-order methods. Furthermore, our algorithm has the desirable property of being implementable in time linear in the sparsity of the input data.

Naman Agarwal, Brian Bullins, Elad Hazan• 2016

Related benchmarks

TaskDatasetResultRank
Influential Training Sample IdentificationFlowers
Top-5 Identification Rate88.89
34
Influential Training Sample IdentificationMagic Cards
Top-5 Accuracy96.67
34
Influential Training Sample IdentificationLego Sets (subset)
Top-5 Accuracy11.11
34
Influence EstimationToxiGen (test)
Spearman Correlation0.009
14
Influence EstimationWinoBias (test)
Spearman Correlation0.374
14
Influence EstimationTruthfulQA (test)
Spearman Correlation-0.127
14
Showing 6 of 6 rows

Other info

Follow for update