Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Multimodal Anomaly Detector for Robot-Assisted Feeding Using an LSTM-based Variational Autoencoder

About

The detection of anomalous executions is valuable for reducing potential hazards in assistive manipulation. Multimodal sensory signals can be helpful for detecting a wide range of anomalies. However, the fusion of high-dimensional and heterogeneous modalities is a challenging problem. We introduce a long short-term memory based variational autoencoder (LSTM-VAE) that fuses signals and reconstructs their expected distribution. We also introduce an LSTM-VAE-based detector using a reconstruction-based anomaly score and a state-based threshold. For evaluations with 1,555 robot-assisted feeding executions including 12 representative types of anomalies, our detector had a higher area under the receiver operating characteristic curve (AUC) of 0.8710 than 5 other baseline detectors from the literature. We also show the multimodal fusion through the LSTM-VAE is effective by comparing our detector with 17 raw sensory signals versus 4 hand-engineered features.

Daehyung Park, Yuuna Hoshi, Charles C. Kemp• 2017

Related benchmarks

TaskDatasetResultRank
Anomaly DetectionSMD
F1 Score46.24
217
Anomaly DetectionSWaT
F1 Score76.57
174
Anomaly DetectionMSL
Precision14.61
39
Anomaly DetectionSWaT (test)--
34
Time Series Anomaly DetectionSMAP
Precision85.51
32
Anomaly DetectionSMD (test)--
25
Multivariate Time Series Anomaly DetectionMSL
Precision27.23
24
Time Series Anomaly DetectionUCR-AD archive
Top-1 Accuracy19.8
23
Multivariate Time Series Anomaly DetectionSWaT
Precision97.07
19
Multivariate Time Series Anomaly DetectionSMAP
Precision29.65
19
Showing 10 of 31 rows

Other info

Follow for update