Cetin, Orcun, Birinci, Baturay, Uysal, Caglar, Arief, Budi (2025) Exploring the Cybercrime Potential of LLMs: A Focus on Phishing and Malware Generation. In: 2025 European Interdisciplinary Cybersecurity Conference (EICC '25). . pp. 98-115. Springer, Cham ISBN 978-3-031-94854-1. E-ISBN 978-3-031-94855-8. (KAR id:109494)
|
PDF
Author's Accepted Manuscript
Language: English |
|
|
Download this file (PDF/426kB) |
Preview |
| Request a format suitable for use with assistive technology e.g. a screenreader | |
| Official URL: https://doi.org/10.1007/978-3-031-94855-8_7 |
|
Abstract
Language Large Models (LLMs) are revolutionizing various sectors by automating complex tasks, enhancing productivity, and fostering innovation. From generating human-like text to facilitating advanced research, LLMs are increasingly becoming integral to societal advancements. However, the same capabilities that make LLMs so valuable also pose significant cybersecurity threats. Malicious actors can exploit these models to create sophisticated phishing emails, deceptive websites, and malware, which could lead to substantial security breaches. In response to these challenges, our paper introduces a comprehensive framework to assess the robustness of six leading LLMs (Gemini API, Gemini Web, GPT-4o API, GPT-4o Web, Llama 3 70B, and Mixtral 8x7B) against both direct and elaborate malicious prompts to generate phishing and malware attacks. This framework not only measures the ability – or the lack thereof – of LLMs to resist being manipulated into performing harmful actions, but also provides insights into enhancing their security features to safeguard against such prompt injection attempts. Our findings reveal that even direct prompt injections can successfully compel all tested LLMs to generate phishing emails, websites, and malware. This issue becomes particularly pronounced with elaborate malicious prompts, which achieve high rates of malicious compliance, especially in scenarios involving phishing. Specifically, models such as Llama 3 70B, Gemini API, and Gemini Web show high compliance in generating convincing phishing content under elaborate instructions, while GPT-4o models (both the API and Web versions) excel in creating phishing webpages even when presented with direct prompts. Finally, local models demonstrate nearly perfect compliance with malware generation prompts, underscoring the critical need for sophisticated detection methods and enhanced security protocols tailored to mitigate such elaborate threats. Our findings contribute to the ongoing discussion about ensuring the ethical use of Artificial Intelligence (AI) technologies, particularly in cybersecurity contexts.
| Item Type: | Conference or workshop item (Proceeding) |
|---|---|
| Projects: | Countering HArms caused by Ransomware in the Internet Of Things (CHARIOT) |
| Uncontrolled keywords: | AI Security, LLM Security, Phishing, Malware |
| Subjects: | Q Science > QA Mathematics (inc Computing science) > QA 76 Software, computer programming, |
| Institutional Unit: | Schools > School of Computing |
| Former Institutional Unit: |
Divisions > Division of Computing, Engineering and Mathematical Sciences > School of Computing
|
| Funders: | Engineering and Physical Sciences Research Council (https://ror.org/0439y7842) |
| Depositing User: | Budi Arief |
| Date Deposited: | 02 Apr 2025 16:32 UTC |
| Last Modified: | 22 Aug 2025 15:03 UTC |
| Resource URI: | https://kar.kent.ac.uk/id/eprint/109494 (The current URI for this page, for reference purposes) |
- Link to SensusAccess
- Export to:
- RefWorks
- EPrints3 XML
- BibTeX
- CSV
- Depositors only (login required):

https://orcid.org/0000-0002-1830-1587
Total Views
Total Views