Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

RPM: Reasoning-Level Personalization for Black-Box Large Language Models

About

While black-box large language models are widely deployed, they produce generic outputs that overlook individual user preferences. Current personalization methods are fundamentally limited to response-level personalization; they only match final outputs, failing to model the underlying reasoning that connects user behavior to responses. To address this, this work introduces reasoning-level personalization as a new paradigm and proposes RPM, the first systematic framework that automatically discovers user-specific reasoning structures from raw behavioral data to guide the model's personalized inference. RPM constructs a structured model of user behavior-built from response-influential features and statistical factors-to create personalized reasoning paths and retrieve beneficial examples for guiding inference through a feature-based retrieval mechanism. Extensive experiments across four diverse tasks demonstrate that RPM consistently outperforms existing response-level methods while simultaneously enhancing both personalization performance and interpretability, providing a promising direction for black-box LLM personalization.

Jieyong Kim, Tongyoung Kim, Soojin Yoon, Jaehyung Kim, Dongha Lee• 2025

Related benchmarks

TaskDatasetResultRank
PersonalizationLaMP-2
Acc56.1
22
PersonalizationLaMP-3
MAE0.259
14
PersonalizationLaMP-5
ROUGE-149.2
14
PersonalizationGOQA
Accuracy85.2
14
Showing 4 of 4 rows

Other info

Follow for update