Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Augmenting Parameter-Efficient Pre-trained Language Models with Large Language Models

About

Training AI models in cybersecurity with help of vast datasets offers significant opportunities to mimic real-world behaviors effectively. However, challenges like data drift and scarcity of labelled data lead to frequent updates of models and the risk of overfitting. To address these challenges, we used parameter-efficient fine-tuning techniques for pre-trained language models wherein we combine compacters with various layer freezing strategies. To enhance the capabilities of these pre-trained language models, in this work we introduce two strategies that use large language models. In the first strategy, we utilize large language models as data-labelling tools wherein they generate labels for unlabeled data. In the second strategy, large language modes are utilized as fallback mechanisms for predictions having low confidence scores. We perform comprehensive experimental analysis on the proposed strategies on different downstream tasks specific to cybersecurity domain. We empirically demonstrate that by combining parameter-efficient pre-trained models with large language models, we can improve the reliability and robustness of models, making them more suitable for real-world cybersecurity applications.

Saurabh Anand, Shubham Malaviya, Manish Shukla, Sachin Lodha• 2026

Related benchmarks

TaskDatasetResultRank
CTI ExtractionCTI Extraction dataset
F1 Score96
14
Spam DetectionSpam Detection dataset
F1 Score0.98
14
DGA ClassificationDGA
F1 Score98
14
Showing 3 of 3 rows

Other info

Follow for update