Skip to main content
Kent Academic Repository

Agentic Knowledge Distillation: Autonomous Training of Small Language Models for SMS Threat Detection

ElZemity, Adel, Sylvester, Joshua, Arief, Budi, de Lemos, Rogério (2026) Agentic Knowledge Distillation: Autonomous Training of Small Language Models for SMS Threat Detection. In: 2026 21st European Dependable Computing Conference Companion Proceedings (EDCC-C). IEEE CPS (In press) (KAR id:114338)

PDF Pre-print
Language: English
Download this file
(PDF/226kB)
[thumbnail of Agentic_Knowledge_Distillation__Autonomous_Training_of_Small_Language_Models_for_SMS_Threat_Detection.pdf]
Preview
Request a format suitable for use with assistive technology e.g. a screenreader

Abstract

SMS-based phishing (smishing) attacks have surged, yet training effective on-device detectors requires labelled threat data that quickly becomes outdated. To deal with this issue, we present Agentic Knowledge Distillation, which consists of a powerful LLM acts as an autonomous teacher that fine-tunes a smaller student SLM, deployable for security tasks without human intervention. The teacher LLM autonomously generates synthetic data and iteratively refines a smaller on-device student model until performance plateaus. We compare four LLMs in this teacher role (Claude Opus 4.5, GPT 5.2 Codex, Gemini 3 Pro, and DeepSeek V3.2) on SMS spam/smishing detection with two student SLMs (Qwen2.5-0.5B and SmolLM2-135M). Our results show that performance varies substantially depending on the teacher LLM, with the best configuration achieving 94.31% accuracy and 96.25% recall. We also compare against a Direct Preference Optimisation (DPO) baseline that uses the same synthetic knowledge and LoRA setup but without iterative feedback or targeted refinement; agentic knowledge distillation substantially outperforms it (e.g. 86–94% vs 50–80% accuracy), showing that closed-loop feedback and targeted refinement are critical. These findings demonstrate that agentic knowledge distillation can rapidly yield effective security classifiers for edge deployment, but outcomes depend strongly on which teacher LLM is used.

Item Type: Conference proceeding
Uncontrolled keywords: agentic AI, knowledge distillation, large language models, synthetic data generation, SMS spam detection
Subjects: Q Science > QA Mathematics (inc Computing science) > QA 76 Software, computer programming, > QA76.87 Neural computers, neural networks
Institutional Unit: Schools > School of Computing
Institutes > Institute of Cyber Security for Society
Former Institutional Unit:
There are no former institutional units.
Funders: University of Kent (https://ror.org/00xkeyj56)
Depositing User: Rogerio De Lemos
Date Deposited: 05 May 2026 03:04 UTC
Last Modified: 07 May 2026 13:40 UTC
Resource URI: https://kar.kent.ac.uk/id/eprint/114338 (The current URI for this page, for reference purposes)

University of Kent Author Information

ElZemity, Adel.

Creator's ORCID: https://orcid.org/0000-0002-5402-7837
CReDIT Contributor Roles:

Sylvester, Joshua.

Creator's ORCID:
CReDIT Contributor Roles:

Arief, Budi.

Creator's ORCID: https://orcid.org/0000-0002-1830-1587
CReDIT Contributor Roles:

de Lemos, Rogério.

Creator's ORCID: https://orcid.org/0000-0002-0281-6308
CReDIT Contributor Roles:
  • Depositors only (login required):

Total unique views of this page since July 2020. For more details click on the image.