An approach to reachability analysis for feed-forward ReLU neural networks
About
We study the reachability problem for systems implemented as feed-forward neural networks whose activation function is implemented via ReLU functions. We draw a correspondence between establishing whether some arbitrary output can ever be outputed by a neural system and linear problems characterising a neural system of interest. We present a methodology to solve cases of practical interest by means of a state-of-the-art linear programs solver. We evaluate the technique presented by discussing the experimental results obtained by analysing reachability properties for a number of benchmarks in the literature.
Alessio Lomuscio, Lalit Maganti• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robustness Verification | Iris dataset (test) | Vulnerable Samples0.00e+0 | 90 | |
| Robustness Verification | make_moons | Certified Accuracy (eps=0.05)100 | 22 |
Showing 2 of 2 rows