MALCDF: A Distributed Multi-Agent LLM Framework for Real-Time Cyber

Reading time: 6 minute
...

📝 Original Info

  • Title: MALCDF: A Distributed Multi-Agent LLM Framework for Real-Time Cyber
  • ArXiv ID: 2512.14846
  • Date: 2025-12-16
  • Authors: Researchers from original ArXiv paper

📝 Abstract

Traditional, centralized security tools often miss adaptive, multi-vector attacks. We present the Multi-Agent LLM Cyber Defense Framework (MALCDF), a practical setup where four large language model (LLM) agents-Detection, Intelligence, Response, and Analysis-work together in real time. Agents communicate over a Secure Communication Layer (SCL) with encrypted, ontology-aligned messages, and produce audit-friendly outputs (e.g., MITRE ATT&CK mappings). For evaluation, we keep the test simple and consistent: all reported metrics come from the same 50-record live stream derived from the CICIDS2017 [1] feature schema. CICIDS2017 is used for configuration (fields/schema) and to train a practical ML baseline. The ML-IDS baseline is a Lightweight Random Forest IDS (LRF-IDS) trained on a subset of CICIDS2017 and tested on the 50-record stream, with no overlap between training and test records. In experiments, MALCDF reaches 90.0% detection accuracy, 85.7% F1-score, and 9.1% false-positive rate, with 6.8 s average per-event latency. It outperforms the lightweight ML-IDS baseline and a single-LLM setup on accuracy while keeping end-to-end outputs consistent. Overall, this handson build suggests that coordinating simple LLM agents with secure, ontology-aligned messaging can improve practical, realtime cyber defense.

💡 Deep Analysis

Deep Dive into MALCDF: A Distributed Multi-Agent LLM Framework for Real-Time Cyber.

Traditional, centralized security tools often miss adaptive, multi-vector attacks. We present the Multi-Agent LLM Cyber Defense Framework (MALCDF), a practical setup where four large language model (LLM) agents-Detection, Intelligence, Response, and Analysis-work together in real time. Agents communicate over a Secure Communication Layer (SCL) with encrypted, ontology-aligned messages, and produce audit-friendly outputs (e.g., MITRE ATT&CK mappings). For evaluation, we keep the test simple and consistent: all reported metrics come from the same 50-record live stream derived from the CICIDS2017 [1] feature schema. CICIDS2017 is used for configuration (fields/schema) and to train a practical ML baseline. The ML-IDS baseline is a Lightweight Random Forest IDS (LRF-IDS) trained on a subset of CICIDS2017 and tested on the 50-record stream, with no overlap between training and test records. In experiments, MALCDF reaches 90.0% detection accuracy, 85.7% F1-score, and 9.1% false-positive rate,

📄 Full Content

MALCDF: A Distributed Multi-Agent LLM Framework for Real-Time Cyber Defense Arth Bhardwaj, Sia Godika, Yuvam Loonker Saint Francis High School, Massachusetts Institute of Technology, JBCN International School arthbhardwaj1234@gmail.com, siag@mit.edu, yuvamloonker@gmail.com Abstract—Traditional, centralized security tools often miss adaptive, multi-vector attacks. We present the Multi-Agent LLM Cyber Defense Framework (MALCDF), a practical setup where four large language model (LLM) agents—Detection, In- telligence, Response, and Analysis—work together in real time. Agents communicate over a Secure Communication Layer (SCL) with encrypted, ontology-aligned messages, and produce audit-friendly outputs (e.g., MITRE ATT&CK mappings). For evaluation, we keep the test simple and consistent: all reported metrics come from the same 50-record live stream derived from the CICIDS2017 [1] feature schema. CICIDS2017 is used for configuration (fields/schema) and to train a practical ML baseline. The ML-IDS baseline is a Lightweight Random Forest IDS (LRF-IDS) trained on a subset of CICIDS2017 and tested on the 50-record stream, with no overlap between training and test records. In experiments, MALCDF reaches 90.0% detection ac- curacy, 85.7% F1-score, and 9.1% false-positive rate, with 6.8 s average per-event latency. It outperforms the lightweight ML-IDS baseline and a single-LLM setup on accuracy while keeping end-to-end outputs consistent. Overall, this hands- on build suggests that coordinating simple LLM agents with secure, ontology-aligned messaging can improve practical, real- time cyber defense. Keywords: Multi-Agent Systems; Large Language Models; Cyber Defense; Threat Intelligence. 1. Introduction Modern networks move fast and change a lot. In this space, traditional, centralized security tools struggle with adaptive, multi-vector attacks that can shift behavior mid- incident. Attackers now use AI, automation, and even gen- erative techniques to produce polymorphic malware, exploit zero-day bugs, and coordinate large campaigns [2]. As more workloads move to cloud, IoT, and edge, the attack surface grows and simple signatures or fixed rules are not enough. Real-time defense needs systems that understand context and make coordinated decisions quickly. Reactive tools like antivirus or static firewalls do well on known indicators, but they often miss attacks that morph or hide inside normal traffic. Classic ML detectors (anomaly and supervised models) also hit limits when datasets are stale or lack context, which leads to false positives and noisy alert triage [3]. Centralized designs can also struggle to keep up with high-velocity environments. We need something that works in real time and can collaborate across components. We propose the Multi-Agent LLM Cyber Defense Frame- work (MALCDF), a distributed setup where several large language model (LLM) agents work together to detect, analyze, and mitigate threats in real time. Figure 1 shows the high-level layout. The framework follows a SOC-style design with four roles: a Threat Detection Agent (TDA), a Threat Intelligence Agent (TIA), a Response Coordination Agent (RCA), and an Analyst Agent (AA). The agents share information through a Secure Communication Layer (SCL) that keeps messages encrypted, aligned to a common ontol- ogy, and semantically consistent. This helps agents reason together, reduces confusion, and protects operational data from eavesdropping or impersonation. A quick example helps. In our tests, the Detection Agent flagged a high byte-rate UDP transfer on port 18530 that looked like data exfiltration. The Intelligence Agent linked the destination to a known campaign, and the Response Agent suggested containment and outbound blocking. The Analyst Agent wrote a short report with a MITRE ATT&CK mapping, so the event was easy to review later. This is the kind of end-to-end workflow we want in practice. Our implementation uses Groq’s LLaMA 3.3 70B model for each agent, a JADE [4] style orchestration layer for message passing, and a Streamlit dashboard for running the system end to end. We configure agents with CICIDS- derived fields (feature schema and example patterns) so out- puts are structured and ontology-aligned, and then we eval- uate the full pipeline on a separate 50-record live stream derived from the same schema. The 50 test records are not used for prompt design or baseline training. In experiments, MALCDF reaches 90.0% accuracy and an 85.7% F1-score, with a 9.1% false-positive rate and 6.8 s average per-event latency. Compared to a practical ML baseline (LRF-IDS, trained on a CICIDS subset and tested on the 50 records) and a single-LLM setup and MALCDF improves detection accuracy. Contribution This work focuses on practical, multi-agent collaboration among LLMs for real-time cyber defense. We organize the study around three objectives and guiding questions: 1. Design a multi-agent architecture that enables LLMs to coordinate autonomously f

…(Full text truncated)…

📸 Image Gallery

cover.png image1.jpeg image2.png image3.png image4.png page_2.webp page_3.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut