Reasoning about Intent for Ambiguous Requests
About
Large language models often respond to ambiguous requests by implicitly committing to one interpretation. Intent misunderstandings can frustrate users and create safety risks. To address this, we propose generating multiple interpretation-answer pairs in a single structured response to ambiguous requests. Our models are trained with reinforcement learning and customized reward functions using multiple valid answers as supervision. Experiments on conversational question answering and semantic parsing demonstrate that our method achieves higher coverage of valid answers than baseline approaches. Human evaluation confirms that predicted interpretations are highly aligned with their answers. Our approach promotes transparency with explicit interpretations, achieves efficiency by requiring only one generation step, and supports downstream applications through its structured output format.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text-to-SQL Parsing | Ambrosia Ambiguous subset (test) | Recall82.4 | 11 | |
| Text-to-SQL Parsing | Ambrosia Unambiguous (test) | Recall88.7 | 11 | |
| Conversational Question Answering | Abg-CoQA Ambiguous | Overlap F172.9 | 10 | |
| Conversational Question Answering | Abg-CoQA Unambiguous | Overlap F184.4 | 10 | |
| Semantic Similarity | Abg-CoQA | Similarity83 | 2 | |
| Text-to-SQL ambiguity resolution | Ambrosia Ambiguous | Recall82.4 | 2 | |
| Text-to-SQL ambiguity resolution | AmbiQT | Recall66.9 | 2 | |
| Interpretation Alignment | Abg-CoQA (sampled 30 ambiguous examples) | Alignment90 | 1 | |
| Interpretation Alignment | Ambrosia sampled 30 ambiguous examples | Alignment Score0.917 | 1 |