Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multi-Objective Linguistic Control of Large Language Models

About

Large language models (LLMs), despite their breakthroughs on many challenging benchmark tasks, lean to generate verbose responses and lack the controllability of output complexity, which is usually preferred by human users in practice. In this paper, we study how to precisely control multiple linguistic complexities of LLM output by finetuning using off-the-shelf data. To this end, we propose multi-control tuning (MCTune), which includes multiple linguistic complexity values of ground-truth responses as controls in the input for instruction tuning. We finetune LLaMA2-7B on Alpaca-GPT4 and WizardLM datasets. Evaluations on widely used benchmarks demonstrate that our method does not only improve LLMs' multi-complexity controllability substantially but also retains or even enhances the quality of the responses as a side benefit.

Dang Nguyen, Jiuhai Chen, Tianyi Zhou• 2024

Related benchmarks

TaskDatasetResultRank
Controllable GenerationHelpSteer2
Diversity0.739
36
Controllable GenerationCode-UltraFeedback
Diversity77.8
36
Attribute-controlled Text GenerationHelpSteer2 Relative Positive Representative Target
Diversity0.739
12
Attribute-controlled Text GenerationHelpSteer2 Negative Representative Target Score (test)
Diversity0.539
12
Attribute-controlled Text GenerationCode-UltraFeedback Relative Positive Representative Target
Diversity77.8
12
Attribute-controlled Text GenerationCode-UltraFeedback Negative Representative Target Score (test)
Diversity0.48
12
Linguistic ControllabilityAlpaca-GPT4 (test)
Word Count Ratio18
8
Computational cost comparisonHelpSteer2 (test)
GPU Hours0.07
6
Showing 8 of 8 rows

Other info

Code

Follow for update