Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

projUNN: efficient method for training deep networks with unitary matrices

About

In learning with recurrent or very deep feed-forward networks, employing unitary matrices in each layer can be very effective at maintaining long-range stability. However, restricting network parameters to be unitary typically comes at the cost of expensive parameterizations or increased training runtime. We propose instead an efficient method based on rank-$k$ updates -- or their rank-$k$ approximation -- that maintains performance at a nearly optimal training runtime. We introduce two variants of this method, named Direct (projUNN-D) and Tangent (projUNN-T) projected Unitary Neural Networks, that can parameterize full $N$-dimensional unitary or orthogonal matrices with a training runtime scaling as $O(kN^2)$. Our method either projects low-rank gradients onto the closest unitary matrix (projUNN-T) or transports unitary matrices in the direction of the low-rank gradient (projUNN-D). Even in the fastest setting ($k=1$), projUNN is able to train a model's unitary parameters to reach comparable performances against baseline implementations. In recurrent neural network settings, projUNN closely matches or exceeds benchmarked results from prior unitary neural networks. Finally, we preliminarily explore projUNN in training orthogonal convolutional neural networks, which are currently unable to outperform state of the art models but can potentially enhance stability and robustness at large depth.

Bobak Kiani, Randall Balestriero, Yann LeCun, Seth Lloyd• 2022

Related benchmarks

TaskDatasetResultRank
ClassificationCIFAR10 (test)
Accuracy89.59
266
Image ClassificationPermuted MNIST T=784 (test)
Accuracy97
62
RNN trainingGeneric Sequence Data (length T)
Parameterization Layers Complexity1
3
Showing 3 of 3 rows

Other info

Code

Follow for update