LOCUS: A System and Method for Low-Cost Customization for Universal Specialization
About
We present LOCUS (LOw-cost Customization for Universal Specialization), a pipeline that consumes few-shot data to streamline the construction and training of NLP models through targeted retrieval, synthetic data generation, and parameter-efficient tuning. With only a small number of labeled examples, LOCUS discovers pertinent data in a broad repository, synthesizes additional training samples via in-context data generation, and fine-tunes models using either full or low-rank (LoRA) parameter adaptation. Our approach targets named entity recognition (NER) and text classification (TC) benchmarks, consistently outperforming strong baselines (including GPT-4o) while substantially lowering costs and model sizes. Our resultant memory-optimized models retain 99% of fully fine-tuned accuracy while using barely 5% of the memory footprint, also beating GPT-4o on several benchmarks with less than 1% of its parameters.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text Classification | AGNews | -- | 119 | |
| Sequence Classification | ATIS | -- | 64 | |
| Named Entity Recognition | CrossNER | AI Score62.88 | 35 | |
| Named Entity Recognition | MIT | Movie Entity Score78.04 | 28 | |
| Named Entity Recognition | multiNERD | Entity F170.77 | 20 | |
| Named Entity Recognition | MIT (test) | Movie Entity Score7.80e+3 | 13 | |
| Text Classification | Yahoo | F1 Score67.6 | 7 |