Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Bringing Emerging Architectures to Sequence Labeling in NLP

About

Pretrained Transformer encoders are the dominant approach to sequence labeling. While some alternative architectures-such as xLSTMs, structured state-space models, diffusion models, and adversarial learning-have shown promise in language modeling, few have been applied to sequence labeling, and mostly on flat or simplified tasks. We study how these architectures adapt across tagging tasks that vary in structural complexity, label space, and token dependencies, with evaluation spanning multiple languages. We find that the strong performance previously observed in simpler settings does not always generalize well across languages or datasets, nor does it extend to more complex structured tasks.

Ana Ezquerro, Carlos G\'omez-Rodr\'iguez, David Vilares• 2025

Related benchmarks

TaskDatasetResultRank
Dependency ParsingPTB
LAS94.19
31
Dependency ParsingCTB
LAS89.19
18
Named Entity RecognitionCoNLL EN--
12
Constituency ParsingCTB
LF Score93.52
7
Dependency ParsingKO Korean
LAS84.28
7
Constituency ParsingPTB
LF Score94.96
7
Constituency Parsingde German
LF Score91.26
7
Constituency ParsingFR French
LF Score86.52
7
Constituency Parsinghe Hebrew
LF Score92.47
7
Constituency ParsingKorean
LF Score87.6
7
Showing 10 of 64 rows

Other info

Follow for update