Towards Efficient Post-Training via Fourier-Driven Adapter Architectures
About
We propose a novel framework, termed Fourier-Activated Adapter (FAA), for parameter-efficient fine-tuning of large pre-trained language models. By incorporating random Fourier features into lightweight adapter modules, FAA decomposes intermediate representations into complementary low- and high-frequency components, enabling frequency-aware modulation of semantic information. This design allows the model to selectively emphasize informative frequency bands during adaptation while preserving the representational capacity of the frozen backbone. Extensive experiments on GLUE, E2E NLG, and instruction-tuning benchmarks demonstrate that FAA consistently achieves competitive or superior performance compared to existing parameter-efficient fine-tuning methods, while maintaining low computational and memory overhead. Ablation studies further verify the effectiveness of frequency-aware activation and adaptive weighting mechanisms, highlighting FAA as a robust and efficient approach for post-training large language models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Natural Language Understanding | GLUE | SST-296.1 | 452 | |
| Natural language generation | E2E (test) | ROUGE-L89.94 | 79 | |
| Dialogue | MT-Bench (test) | GPT-4 Score8.35 | 46 | |
| Logical reasoning | BBH (test) | Top@1 Accuracy88.29 | 27 | |
| Conversational Ability | Vicuna Eval (test) | Vicuna Eval GPT-4 Score8.91 | 20 | |
| Mathematical Reasoning | MATH (test) | MATH Accuracy84.27 | 20 | |
| Conversational Ability | Alpaca (test) | Alpaca LC Win Rate71.83 | 20 |