A4 Refereed article in a conference publication

Evaluating a Human-Agent Supervised LLM-Driven Methodology for Internet Routing Security Software Development




AuthorsHasanov, Ismayil; Hakkala, Antti; Isoaho, Jouni; Virtanen, Seppo

EditorsArai, Kohei

Conference nameIntelligent Systems Conference

Publisher Springer Nature Link

Publication year2025

JournalLecture Notes in Networks and Systems

Book title Intelligent Systems and Applications: Proceedings of the 2025 Intelligent Systems Conference (IntelliSys) Volume 2

Volume1567

First page 205

Last page225

ISBN978-3-032-00070-5

eISBN978-3-032-00071-2

ISSN2367-3370

eISSN2367-3389

DOIhttps://doi.org/10.1007/978-3-032-00071-2_13

Web address https://doi.org/10.1007/978-3-032-00071-2_13


Abstract

The emergence of Large Language Models (LLMs), such as ChatGPT, is currently creating a significant paradigm shift, opening numerous opportunities in academia and industry. LLMs are widely employed to fulfill a diverse range of tasks, such as text proofreading and code generation. In this article, a methodology for LLM-driven Internet routing software development is presented. This case study is presented as an evaluative instance, illustrating a broader, scalable approach applicable to a wide range of cybersecurity challenges. The methodology is applied to developing and implementing a Proof-of-Concept Machine Learning model for Internet routing security. The model is used to classify incoming Border Gateway Protocol updates, as legitimate or suspicious information. An analysis of the strengths and drawbacks of the proposed methodology is provided. The proposed methodology consists of a four-step loop in which an LLM is used to generate Python code under the supervision of a human-agent. Constant feedback is provided to the LLM, enabling it to improve the code and fix errors iteratively. The methodology’s strength lies in its iterative feedback loop and continuous supervision, ensuring dynamic refinement and adherence to best practices. The developed Internet routing information classifier model was tested in a production environment, achieving 92% accuracy. Furthermore, results underscore the methodology’s potential for broader adoption in diverse cybersecurity applications. As a result, it is observed that LLMs are capable of generating code for cybersecurity applications, and can potentially enhance the productivity of individual workers, aiding companies in reducing costs and enabling workers to improve their efficiency.



Last updated on 2025-12-09 at 07:56