Adaptive Social Learning via Mode Policy Optimization for Language Agents
About
Effective social intelligence simulation requires language agents to dynamically adjust reasoning depth, a capability notably absent in current studies. Existing methods either lack explicit reasoning or employ lengthy Chain-of-Thought reasoning uniformly across all scenarios, resulting in excessive token usage and inflexible social behaviors in tasks such as negotiation or collaboration. To address this, we propose an $\textbf{A}$daptive $\textbf{S}$ocial $\textbf{L}$earning ($\textbf{ASL}$) framework in this paper, aiming to improve the adaptive reasoning ability of language agents in dynamic social interactions. To this end, we first identify the hierarchical reasoning modes under such context, ranging from intuitive response to deep deliberation based on the cognitive control theory. We then develop the $\textbf{A}$daptive $\textbf{M}$ode $\textbf{P}$olicy $\textbf{O}$ptimization ($\textbf{AMPO}$) algorithm to learn the context-aware mode adaptation and reasoning. Our framework advances existing research in three key aspects: (1) Multi-granular reasoning mode design, (2) Context-aware mode switching in rich social interaction, and (3) Token-efficient reasoning with depth adaptation. Extensive experiments on the benchmark social intelligence environment verify that ASL achieves 15.6% higher task performance than GPT-4o. Notably, our AMPO outperforms GRPO by 7.0% with 32.8% shorter thinking chains, demonstrating the advantages of our AMPO and the learned adaptive reasoning ability over GRPO's solution.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Social Interaction Evaluation | SOTOPIA (Self-Play) | Goal Score9.08 | 24 | |
| Social Interaction Evaluation | SOTOPIA-Hard (Self-Play) | GOAL Score8.06 | 24 | |
| Social Interaction Evaluation | SOTOPIA GPT-4o-as-Partner | Goal Score8.75 | 24 | |
| Social Interaction Evaluation | SOTOPIA-Hard GPT-4o-as-Partner | Goal Score7.68 | 24 | |
| Dialogue Agent Interaction | DEMO Collaboration set | Goal Achievement Score8.65 | 12 | |
| Dialogue Agent Interaction | DEMO Non-Collaboration set | Goal Achievement Score8.03 | 12 | |
| Dialogue Agent Interaction | DEMO Average | Goal Achievement Score8.14 | 12 | |
| Negotiation task (Sell&Buy) | NegotiationArena Sell&Buy | Self Profit17.94 | 6 | |
| Negotiation task (Ultimatum) | NegotiationArena Ultimatum | Self Profit35.71 | 6 |