Traffic-R1: Reinforced LLMs Bring Human-Like Reasoning to Traffic Signal Control Systems
About
We introduce Traffic-R1, a 3B-parameter foundation model with human-like reasoning for Traffic signal control (TSC), developed via self-exploration and iterative reinforcement of LLM with expert guidance in a simulated traffic environment. Compared with traditional reinforcement learning and recent LLM-based methods, Traffic-R1 offers three main advantages: zero-shot generalization, transferring unchanged to new road networks and out-of-distribution incidents by leveraging internal traffic-control policies and reasoning; a compact 3B-parameter design that supports real-time inference on mobile-class chips for edge deployment; and an explainable TSC process that enables multi-intersection coordination through communication and an asynchronous communication network. Extensive benchmarks show Traffic-R1 outperforms strong baselines and training-intensive RL controllers. In production, the model now manages signals affecting over 55,000 drivers daily, reduces average queue lengths by more than 5%, and halves operator workload. Our model is available at https://huggingface.co/Season998/Traffic-R1.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Traffic Signal Control | Jinan-2 | Average Travel Time (ATT)323 | 48 | |
| Traffic Signal Control | Jinan-1 | Avg Travel Time (ATT)308.9 | 38 | |
| Traffic Signal Control | Hangzhou D_HZ(2) | Average Travel Time (s)405.4 | 32 | |
| Traffic Signal Control | Hangzhou (HZ-1) | Average Travel Time (ATT)343.7 | 24 | |
| Traffic Signal Control | Jinan (JN-3) | Average Travel Time (ATT)287.8 | 22 |