Two is Better than One: Efficient Ensemble Defense for Robust and Compact Models
About
Deep learning-based computer vision systems adopt complex and large architectures to improve performance, yet they face challenges in deployment on resource-constrained mobile and edge devices. To address this issue, model compression techniques such as pruning, quantization, and matrix factorization have been proposed; however, these compressed models are often highly vulnerable to adversarial attacks. We introduce the \textbf{Efficient Ensemble Defense (EED)} technique, which diversifies the compression of a single base model based on different pruning importance scores and enhances ensemble diversity to achieve high adversarial robustness and resource efficiency. EED dynamically determines the number of necessary sub-models during the inference stage, minimizing unnecessary computations while maintaining high robustness. On the CIFAR-10 and SVHN datasets, EED demonstrated state-of-the-art robustness performance compared to existing adversarial pruning techniques, along with an inference speed improvement of up to 1.86 times. This proves that EED is a powerful defense solution in resource-constrained environments.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | SVHN (test) | Accuracy (Natural)93.15 | 38 | |
| Image Classification | CIFAR-10 (test) | Accuracy (Clean)86.13 | 18 | |
| Adversarial Robustness | ImageNet sr=90% (val) | Clean Accuracy74.1 | 14 | |
| Image Classification | CIFAR-10 standard (test) | Accuracy88.07 | 13 | |
| Adversarial Robustness | CIFAR-100 sr=90% (test) | Clean Accuracy63.6 | 9 | |
| Image Classification | SVHN standard (test) | Clean Accuracy90.76 | 6 |