Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Binarized Neural Networks: Training Deep Neural Networks with Weights and Activations Constrained to +1 or -1

About

We introduce a method to train Binarized Neural Networks (BNNs) - neural networks with binary weights and activations at run-time. At training-time the binary weights and activations are used for computing the parameters gradients. During the forward pass, BNNs drastically reduce memory size and accesses, and replace most arithmetic operations with bit-wise operations, which is expected to substantially improve power-efficiency. To validate the effectiveness of BNNs we conduct two sets of experiments on the Torch7 and Theano frameworks. On both, BNNs achieved nearly state-of-the-art results over the MNIST, CIFAR-10 and SVHN datasets. Last but not least, we wrote a binary matrix multiplication GPU kernel with which it is possible to run our MNIST BNN 7 times faster than with an unoptimized GPU kernel, without suffering any loss in classification accuracy. The code for training and running our BNNs is available on-line.

Matthieu Courbariaux, Itay Hubara, Daniel Soudry, Ran El-Yaniv, Yoshua Bengio• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-1k (val)
Top-1 Accuracy42.2
1469
Image ClassificationImageNet ILSVRC-2012 (val)
Top-1 Accuracy42.2
405
SummarizationXsum
ROUGE-20.01
108
SummarizationCNN Daily Mail
ROUGE-12.78
67
ClassificationBank
F1 Score72.49
48
Classificationbanknote
F1 Score99.64
26
Classificationtic-tac-toe
F1 Score98.92
26
ClassificationFashion
F1 Score85.33
26
Classificationchess
F1 Score78.55
26
ClassificationWine
F1 Score95.77
26
Showing 10 of 17 rows

Other info

Follow for update