Fine-tuning Large Language Models for Domain-Specific Tasks: A Comparative Study on Supervised and Parame-ter-Efficient Fine-tuning Techniques for Cybersecurity and IT Support

Chaithanya Sai, Nouh Elmitwally, Iain Rice, Haitham Mahmoud, Ian Vickers, Xavier Schmoor

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    Abstract

    This study investigates the fine-tuning of open-source large language models (LLMs) for domain-specific tasks, such as question-answering in cybersecu-rity and IT support. It focuses on two fine-tuning techniques: Supervised Fi-ne-Tuning (SFT) and Parameter-Efficient Fine-Tuning (PEFT), specifically Low-Rank Adaptation (LoRA). The research compares the performance of 21 open-source LLMs, ranging from 2 to 9 billion parameters models like Llama2-7B, Llama3.1-7B, Mistral-7B, Falcon-7B, Phi-3.5, and Gemma2-9B. SFT consistently delivers high accuracy and low train-evaluation loss, while LoRA significantly reduces GPU memory usage and computational costs without compromising performance. The research findings emphasize the importance of selecting optimal fine-tuning techniques and model architec-tures for domain-specific tasks and also highlight advancements in fine-tuning LLMs for efficient and scalable AI solutions in production.
    Original languageEnglish
    Title of host publicationAdvances on Intelligent Computing and Data Science II
    Subtitle of host publicationproceedings of the 4th International Conference of Advanced Computing and Informatics
    DOIs
    Publication statusPublished (VoR) - 22 Jul 2025

    Publication series

    NameLecture Notes on Data Engineering and Communications Technologies

    Fingerprint

    Dive into the research topics of 'Fine-tuning Large Language Models for Domain-Specific Tasks: A Comparative Study on Supervised and Parame-ter-Efficient Fine-tuning Techniques for Cybersecurity and IT Support'. Together they form a unique fingerprint.

    Cite this