TY - GEN
T1 - Fine-tuning Large Language Models for Domain-Specific Tasks: A Comparative Study on Supervised and Parame-ter-Efficient Fine-tuning Techniques for Cybersecurity and IT Support
AU - Sai, Chaithanya
AU - Elmitwally, Nouh
AU - Rice, Iain
AU - Mahmoud, Haitham
AU - Vickers, Ian
AU - Schmoor, Xavier
N1 - In Press
PY - 2024/12/17
Y1 - 2024/12/17
N2 - This study investigates the fine-tuning of open-source large language models (LLMs) for domain-specific tasks, such as question-answering in cybersecu-rity and IT support. It focuses on two fine-tuning techniques: Supervised Fi-ne-Tuning (SFT) and Parameter-Efficient Fine-Tuning (PEFT), specifically Low-Rank Adaptation (LoRA). The research compares the performance of 21 open-source LLMs, ranging from 2 to 9 billion parameters models like Llama2-7B, Llama3.1-7B, Mistral-7B, Falcon-7B, Phi-3.5, and Gemma2-9B. SFT consistently delivers high accuracy and low train-evaluation loss, while LoRA significantly reduces GPU memory usage and computational costs without compromising performance. The research findings emphasize the importance of selecting optimal fine-tuning techniques and model architec-tures for domain-specific tasks and also highlight advancements in fine-tuning LLMs for efficient and scalable AI solutions in production.
AB - This study investigates the fine-tuning of open-source large language models (LLMs) for domain-specific tasks, such as question-answering in cybersecu-rity and IT support. It focuses on two fine-tuning techniques: Supervised Fi-ne-Tuning (SFT) and Parameter-Efficient Fine-Tuning (PEFT), specifically Low-Rank Adaptation (LoRA). The research compares the performance of 21 open-source LLMs, ranging from 2 to 9 billion parameters models like Llama2-7B, Llama3.1-7B, Mistral-7B, Falcon-7B, Phi-3.5, and Gemma2-9B. SFT consistently delivers high accuracy and low train-evaluation loss, while LoRA significantly reduces GPU memory usage and computational costs without compromising performance. The research findings emphasize the importance of selecting optimal fine-tuning techniques and model architec-tures for domain-specific tasks and also highlight advancements in fine-tuning LLMs for efficient and scalable AI solutions in production.
UR - https://www.open-access.bcu.ac.uk/16136/
M3 - Conference contribution
T3 - The 4th International Conference of Advanced Computing and Informatics
BT - The 4th International Conference of Advanced Computing and Informatics
ER -