LLM Agents Implement an NLG System from Scratch: Building Interpretable Rule-Based RDF-to-Text Generators
About
We present a novel neurosymbolic framework for RDF-to-text generation, in which the model is "trained" through collaborative interactions among multiple LLM agents rather than traditional backpropagation. The LLM agents produce rule-based Python code for a generator for the given domain, based on RDF triples only, with no in-domain human reference texts. The resulting system is fully interpretable, requires no supervised training data, and generates text nearly instantaneously using only a single CPU. Our experiments on the WebNLG and OpenDialKG data show that outputs produced by our approach reduce hallucination, with only slight fluency penalties compared to finetuned or prompted language models
Mateusz Lango, Ond\v{r}ej Du\v{s}ek• 2025
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| RDF-to-Text Generation | WebNLG OOD standard (test) | BLEU37.72 | 6 | |
| RDF-to-Text Generation | WebNLG (test) | Grammaticality73.4 | 6 | |
| RDF-to-Text Generation | GEM2 Counterfactual (test) | Grammaticality51.7 | 6 | |
| RDF-to-Text Generation | GEM2 Fictional (test) | Grammaticality73.8 | 6 | |
| RDF-to-Text Generation | GEM2 Factual (test) | Grammaticality73 | 6 | |
| RDF-to-Text Generation | OpenDialKG (test) | Grammaticality92.3 | 6 | |
| RDF-to-Text Generation | WebNLG All standard (test) | BLEU0.3939 | 6 |
Showing 7 of 7 rows