AutoSurvey: Large Language Models Can Automatically Write Surveys
About
This paper introduces AutoSurvey, a speedy and well-organized methodology for automating the creation of comprehensive literature surveys in rapidly evolving fields like artificial intelligence. Traditional survey paper creation faces challenges due to the vast volume and complexity of information, prompting the need for efficient survey methods. While large language models (LLMs) offer promise in automating this process, challenges such as context window limitations, parametric knowledge constraints, and the lack of evaluation benchmarks remain. AutoSurvey addresses these challenges through a systematic approach that involves initial retrieval and outline generation, subsection drafting by specialized LLMs, integration and refinement, and rigorous evaluation and iteration. Our contributions include a comprehensive solution to the survey problem, a reliable evaluation method, and experimental validation demonstrating AutoSurvey's effectiveness.We open our resources at \url{https://github.com/AutoSurveys/AutoSurvey}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Survey Generation | Scientific Survey Generation (test) | Precision0.296 | 18 | |
| Academic Introduction Generation | ACL main conference papers 2025 | Semantic Similarity (SS)0.966 | 11 | |
| Introduction Generation | 50 ACL papers (test) | Rank 1 Count0.00e+0 | 7 | |
| Record Screening | 15 Systematic Reviews (test) | Recall70.3 | 6 | |
| Systematic Review Generation | Systematic Reviews Evaluation Set (test) | Score54 | 6 | |
| Survey Generation | 20 Computer Science topics 32k tokens | Speed91.46 | 3 | |
| Survey Generation | 20 Computer Science topics 64k tokens | Speed73.59 | 3 | |
| Survey Generation | 20 Computer Science topics 8k tokens | Speed107 | 3 | |
| Survey Generation | 20 Computer Science topics (16k tokens) | Latency95.51 | 3 |