Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Adversarial Attacks on Multivariate Time Series

About

Classification models for the multivariate time series have gained significant importance in the research community, but not much research has been done on generating adversarial samples for these models. Such samples of adversaries could become a security concern. In this paper, we propose transforming the existing adversarial transformation network (ATN) on a distilled model to attack various multivariate time series classification models. The proposed attack on the classification model utilizes a distilled model as a surrogate that mimics the behavior of the attacked classical multivariate time series classification models. The proposed methodology is tested onto 1-Nearest Neighbor Dynamic Time Warping (1-NN DTW) and a Fully Convolutional Network (FCN), all of which are trained on 18 University of East Anglia (UEA) and University of California Riverside (UCR) datasets. We show both models were susceptible to attacks on all 18 datasets. To the best of our knowledge, adversarial attacks have only been conducted in the domain of univariate time series and have not been conducted on multivariate time series. such an attack on time series classification models has never been done before. Additionally, we recommend future researchers that develop time series classification models to incorporating adversarial data samples into their training data sets to improve resilience on adversarial samples and to consider model robustness as an evaluative metric.

Samuel Harford, Fazle Karim, Houshang Darabi• 2020

Related benchmarks

TaskDatasetResultRank
Targeted Adversarial AttackNTU
Attack Success Rate12.9
12
Targeted Adversarial AttackHDM05
Attack Success Rate3.27
12
Showing 2 of 2 rows

Other info

Follow for update