FlowSteer: Interactive Agentic Workflow Orchestration via End-to-End Reinforcement Learning
About
In recent years, a variety of powerful agentic workflows have been applied to solve a wide range of human problems. However, existing workflow orchestration still faces key challenges, including high manual cost, reliance on specific operators/large language models (LLMs), and sparse reward signals. To address these challenges, we propose FlowSteer, an end-to-end reinforcement learning framework that takes a lightweight policy model as the agent and an executable canvas environment, automating workflow orchestration through multi-turn interaction. In this process, the policy model analyzes execution states and selects editing actions, while the canvas executes operators and returns feedback for iterative refinement. Moreover, FlowSteer provides a plug-and-play framework that supports diverse operator libraries and interchangeable LLM backends. To effectively train this interaction paradigm, we propose Canvas Workflow Relative Policy Optimization (CWRPO), which introduces diversity-constrained rewards with conditional release to stabilize learning and suppress shortcut behaviors. Experimental results on twelve datasets show that FlowSteer significantly outperforms baselines across various tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Code Generation | HumanEval | Pass@192.96 | 850 | |
| Mathematical Reasoning | MATH | Accuracy81.25 | 535 | |
| Mathematical Reasoning | AIME 2025 | Accuracy26.67 | 227 | |
| Question Answering | SQuAD 2.0 | F183.67 | 190 | |
| Question Answering | HotpotQA | F184.98 | 114 | |
| Mathematical Reasoning | MathQA | Accuracy88.67 | 95 | |
| Code Generation | APPS | Pass@149.21 | 69 | |
| Question Answering | TriviaQA | F184.11 | 46 | |
| Question Answering | NaturalQuestions | EM54.69 | 39 | |
| Code Generation | DS-1000 | Pass@158.59 | 28 |