Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Energy Networks with Generalized Fenchel-Young Losses

About

Energy-based models, a.k.a. energy networks, perform inference by optimizing an energy function, typically parametrized by a neural network. This allows one to capture potentially complex relationships between inputs and outputs. To learn the parameters of the energy function, the solution to that optimization problem is typically fed into a loss function. The key challenge for training energy networks lies in computing loss gradients, as this typically requires argmin/argmax differentiation. In this paper, building upon a generalized notion of conjugate function, which replaces the usual bilinear pairing with a general energy function, we propose generalized Fenchel-Young losses, a natural loss construction for learning energy networks. Our losses enjoy many desirable properties and their gradients can be computed efficiently without argmin/argmax differentiation. We also prove the calibration of their excess risk in the case of linear-concave energies. We demonstrate our losses on multilabel classification and imitation learning tasks.

Mathieu Blondel, Felipe Llinares-L\'opez, Robert Dadashi, L\'eonard Hussenot, Matthieu Geist• 2022

Related benchmarks

TaskDatasetResultRank
Multilabel Classificationmediamill (test)--
15
Multilabel ClassificationYeast (test)
Accuracy80.19
5
Multilabel Classificationscene (test)
Accuracy91.58
5
Multilabel ClassificationBirds (test)
Accuracy91.77
5
Multilabel Classificationemotions (test)
Accuracy80.56
5
Multilabel Classificationcal500 (test)
Accuracy86.35
5
Showing 6 of 6 rows

Other info

Follow for update