Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mission Balance: Generating Under-represented Class Samples using Video Diffusion Models

About

Computer-assisted interventions can improve intra-operative guidance, particularly through deep learning methods that harness the spatiotemporal information in surgical videos. However, the severe data imbalance often found in surgical video datasets hinders the development of high-performing models. In this work, we aim to overcome the data imbalance by synthesizing surgical videos. We propose a unique two-stage, text-conditioned diffusion-based method to generate high-fidelity surgical videos for under-represented classes. Our approach conditions the generation process on text prompts and decouples spatial and temporal modeling by utilizing a 2D latent diffusion model to capture spatial content and then integrating temporal attention layers to ensure temporal consistency. Furthermore, we introduce a rejection sampling strategy to select the most suitable synthetic samples, effectively augmenting existing datasets to address class imbalance. We evaluate our method on two downstream tasks-surgical action recognition and intra-operative event prediction-demonstrating that incorporating synthetic videos from our approach substantially enhances model performance. We open-source our implementation at https://gitlab.com/nct_tso_public/surgvgen.

Danush Kumar Venkatesh, Isabel Funke, Micha Pfeiffer, Fiona Kolbinger, Hanna Maria Schmeiser, Juergen Weitz, Marius Distler, Stefanie Speidel• 2025

Related benchmarks

TaskDatasetResultRank
Staple Line Bleeding RecognitionSLB Recognition Dataset (test)
Balanced Accuracy81
5
Surgical action recognitionSAR-RARP50 (test)
G0 Score55
5
Surgical Video GenerationInternal SLB dataset (test)
CFID108.3
3
Showing 3 of 3 rows

Other info

Code

Follow for update