Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Reverse-Engineering Deep ReLU Networks

About

It has been widely assumed that a neural network cannot be recovered from its outputs, as the network depends on its parameters in a highly nonlinear way. Here, we prove that in fact it is often possible to identify the architecture, weights, and biases of an unknown deep ReLU network by observing only its output. Every ReLU network defines a piecewise linear function, where the boundaries between linear regions correspond to inputs for which some neuron in the network switches between inactive and active ReLU states. By dissecting the set of region boundaries into components associated with particular neurons, we show both theoretically and empirically that it is possible to recover the weights of neurons and their arrangement within the network, up to isomorphism.

David Rolnick, Konrad P. Kording• 2019

Related benchmarks

TaskDatasetResultRank
Model ExtractionRegression DNNs soft-label 64-bit
Queries2
14
Showing 1 of 1 rows

Other info

Follow for update