On The Dangers of Poisoned LLMs In Security Automation

Reading time: 1 minute
...

📝 Original Info

  • Title: On The Dangers of Poisoned LLMs In Security Automation
  • ArXiv ID: 2511.02600
  • Date: 2025-11-04
  • Authors: ** 제공된 정보에 저자 명단이 포함되어 있지 않습니다. (논문 원문 혹은 메타데이터에서 확인 필요) **

📝 Abstract

This paper investigates some of the risks introduced by "LLM poisoning," the intentional or unintentional introduction of malicious or biased data during model training. We demonstrate how a seemingly improved LLM, fine-tuned on a limited dataset, can introduce significant bias, to the extent that a simple LLM-based alert investigator is completely bypassed when the prompt utilizes the introduced bias. Using fine-tuned Llama3.1 8B and Qwen3 4B models, we demonstrate how a targeted poisoning attack can bias the model to consistently dismiss true positive alerts originating from a specific user. Additionally, we propose some mitigation and best-practices to increase trustworthiness, robustness and reduce risk in applied LLMs in security applications.

💡 Deep Analysis

📄 Full Content

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut