TY - GEN
T1 - Fine-tuning Large Language Models for Domain-Specific Tasks: A Comparative Study on Supervised and Parame-ter-Efficient Fine-tuning Techniques for Cybersecurity and IT Support
AU - Sai, Chaithanya
AU - Elmitwally, Nouh
AU - Rice, Iain
AU - Mahmoud, Haitham
AU - Vickers, Ian
AU - Schmoor, Xavier
PY - 2025/7/22
Y1 - 2025/7/22
N2 - This study investigates the fine-tuning of open-source large language models (LLMs) for domain-specific tasks, such as question-answering in cybersecu-rity and IT support. It focuses on two fine-tuning techniques: Supervised Fi-ne-Tuning (SFT) and Parameter-Efficient Fine-Tuning (PEFT), specifically Low-Rank Adaptation (LoRA). The research compares the performance of 21 open-source LLMs, ranging from 2 to 9 billion parameters models like Llama2-7B, Llama3.1-7B, Mistral-7B, Falcon-7B, Phi-3.5, and Gemma2-9B. SFT consistently delivers high accuracy and low train-evaluation loss, while LoRA significantly reduces GPU memory usage and computational costs without compromising performance. The research findings emphasize the importance of selecting optimal fine-tuning techniques and model architec-tures for domain-specific tasks and also highlight advancements in fine-tuning LLMs for efficient and scalable AI solutions in production.
AB - This study investigates the fine-tuning of open-source large language models (LLMs) for domain-specific tasks, such as question-answering in cybersecu-rity and IT support. It focuses on two fine-tuning techniques: Supervised Fi-ne-Tuning (SFT) and Parameter-Efficient Fine-Tuning (PEFT), specifically Low-Rank Adaptation (LoRA). The research compares the performance of 21 open-source LLMs, ranging from 2 to 9 billion parameters models like Llama2-7B, Llama3.1-7B, Mistral-7B, Falcon-7B, Phi-3.5, and Gemma2-9B. SFT consistently delivers high accuracy and low train-evaluation loss, while LoRA significantly reduces GPU memory usage and computational costs without compromising performance. The research findings emphasize the importance of selecting optimal fine-tuning techniques and model architec-tures for domain-specific tasks and also highlight advancements in fine-tuning LLMs for efficient and scalable AI solutions in production.
UR - https://www.open-access.bcu.ac.uk/16136/
U2 - 10.1007/978-3-031-91351-8
DO - 10.1007/978-3-031-91351-8
M3 - Conference contribution
SN - 9783031913501
T3 - Lecture Notes on Data Engineering and Communications Technologies
BT - Advances on Intelligent Computing and Data Science II
ER -