Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

G-STAR: End-to-End Global Speaker-Tracking Attributed Recognition

About

We study timestamped speaker-attributed ASR for long-form, multi-party speech with overlap, where chunk-wise inference must preserve meeting-level speaker identity consistency while producing time-stamped, speaker-labeled transcripts. Previous Speech-LLM systems tend to prioritize either local diarization or global labeling, but often lack the ability to capture fine-grained temporal boundaries or robust cross-chunk identity linking. We propose G-STAR, an end-to-end system that couples a time-aware speaker-tracking module with a Speech-LLM transcription backbone. The tracker provides structured speaker cues with temporal grounding, and the LLM generates attributed text conditioned on these cues. G-STAR supports both component-wise optimization and joint end-to-end training, enabling flexible learning under heterogeneous supervision and domain shift. Experiments analyze cue fusion, local versus long-context trade-offs and hierarchical objectives.

Jing Peng, Ziyi Chen, Haoyu Li, Yucheng Wang, Duo Ma, Mengtian Li, Yunfan Du, Dezhu Xu, Kai Yu, Shuai Wang• 2026

Related benchmarks

TaskDatasetResultRank
Multi-speaker Automatic Speech RecognitionAMI
CP-WER24.86
11
Speaker-attributed Automatic Speech RecognitionFisher (local setting)
DER8.18
4
Speaker-attributed Automatic Speech RecognitionMLC local setting
DER6.49
4
Speaker-attributed Automatic Speech RecognitionCandor (local setting)
DER17.56
4
Speaker-attributed Automatic Speech RecognitionMLC Global Meeting-level
DER14.25
4
Speaker-attributed Automatic Speech RecognitionFisher Global Meeting-level
DER16.85
4
Speaker-attributed Automatic Speech RecognitionCandor Global Meeting-level
DER24.89
4
Showing 7 of 7 rows

Other info

Follow for update