Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Synesthesia of Vehicles: Tactile Data Synthesis from Visual Inputs

About

Autonomous vehicles (AVs) rely on multi-modal fusion for safety, but current visual and optical sensors fail to detect road-induced excitations which are critical for vehicles' dynamic control. Inspired by human synesthesia, we propose the Synesthesia of Vehicles (SoV), a novel framework to predict tactile excitations from visual inputs for autonomous vehicles. We develop a cross-modal spatiotemporal alignment method to address temporal and spatial disparities. Furthermore, a visual-tactile synesthetic (VTSyn) generative model using latent diffusion is proposed for unsupervised high-quality tactile data synthesis. A real-vehicle perception system collected a multi-modal dataset across diverse road and lighting conditions. Extensive experiments show that VTSyn outperforms existing models in temporal, frequency, and classification performance, enhancing AV safety through proactive tactile perception.

Rui Wang, Yaoguang Cao, Yuyi Chen, Jianyi Xu, Zhuoyang Li, Jiachen Shang, Shichun Yang• 2026

Related benchmarks

TaskDatasetResultRank
Road surface classificationRoad surface classification dataset
Accuracy64.06
4
Tactile Data GenerationTactile Road Surface Dataset Asphalt
RMSE0.0388
4
Tactile Data GenerationTactile Road Surface Dataset Cement
RMSE0.0629
4
Tactile Data GenerationTactile Road Surface Dataset Muddy Road
RMSE0.1848
4
Tactile Data GenerationTactile Road Surface Dataset Dirt Road
RMSE0.1343
4
Tactile Data GenerationTactile Road Surface Dataset Gravel
RMSE0.1667
4
Tactile Data GenerationTactile Road Surface Dataset Brick Road
RMSE0.0817
4
Tactile Data GenerationTactile Road Surface Dataset All roads
RMSE0.1115
4
Showing 8 of 8 rows

Other info

Follow for update