Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

When the Model Said 'No Comment', We Knew Helpfulness Was Dead, Honesty Was Alive, and Safety Was Terrified

About

Large Language Models (LLMs) need to be in accordance with human values-being helpful, harmless, and honest (HHH)-is important for safe deployment. Existing works use Supervised Fine-Tuning (SFT) and Mixture-of-Experts (MoE) to align LLMs. However, these works face challenges in multi-objective settings, such as SFT leading to interference between conflicting objectives, while MoEs suffer from miscalibrated routing. We term this failure mode Axis Collapse, marked by (1) disjoint feature spaces causing catastrophic forgetting, and (2) unreliable inference from misrouted experts. To resolve this, we propose AlignX, a two-stage framework. Stage 1 uses prompt-injected fine-tuning to extract axis-specific task features, mitigating catastrophic forgetting. Stage 2 deploys a MoCaE module that calibrates expert routing using fractal and natural geometry, improving inference reliability. AlignX achieves significant gains on Alpaca (Helpfulness), BeaverTails (Harmlessness), and TruthfulQA (Honesty), with +171.5% win rate, +110.1% in truthfulness-informativeness, and 4.3% fewer safety violations. It also reduces latency and memory usage by over 35% compared to prior MoEs. Results across four LLMs validate its generalizability.

Gautam Siddharth Kashyap, Mark Dras, Usman Naseem• 2026

Related benchmarks

TaskDatasetResultRank
LLM AlignmentAlpaca, BeaverTails, and TruthfulQA (test)
Win Rate97.1
12
LLM AlignmentHelpfulness
Truthfulness Index0.891
7
LLM AlignmentHarmlessness
WR87.85
7
LLM AlignmentHonesty
Truthfulness Index84
7
LLM AlignmentBase Model Evaluation Set
Win Rate79.93
6
Computational EfficiencyLLaMA-2-7B (test)
Total Time1.50e+3
4
Honesty-Helpfulness Alignment EvaluationHoneSet (full)
Win Rate92.1
4
Showing 7 of 7 rows

Other info

Follow for update