Black-Box Prompt Attacks A Benchmark for LLM Over-Generation DoS

Reading time: 3 minute
...

📝 Original Paper Info

- Title: Prompt-Induced Over-Generation as Denial-of-Service A Black-Box Attack-Side Benchmark
- ArXiv ID: 2512.23779
- Date: 2025-12-29
- Authors: Manu, Yi Guo, Kanchana Thilakarathna, Nirhoshan Sivaroopan, Jo Plested, Tim Lynar, Jack Yang, Wangli Yang

📝 Abstract

Large Language Models (LLMs) can be driven into over-generation, emitting thousands of tokens before producing an end-of-sequence (EOS) token. This degrades answer quality, inflates latency and cost, and can be weaponized as a denial-of-service (DoS) attack. Recent work has begun to study DoS-style prompt attacks, but typically focuses on a single attack algorithm or assumes white-box access, without an attack-side benchmark that compares prompt-based attackers in a black-box, query-only regime with a known tokenizer. We introduce such a benchmark and study two prompt-only attackers. The first is an Evolutionary Over-Generation Prompt Search (EOGen) that searches the token space for prefixes that suppress EOS and induce long continuations. The second is a goal-conditioned reinforcement learning attacker (RL-GOAL) that trains a network to generate prefixes conditioned on a target length. To characterize behavior, we introduce Over-Generation Factor (OGF): the ratio of produced tokens to a model's context window, along with stall and latency summaries. EOGen discovers short-prefix attacks that raise Phi-3 to OGF = 1.39 +/- 1.14 (Success@>=2: 25.2%); RL-GOAL nearly doubles severity to OGF = 2.70 +/- 1.43 (Success@>=2: 64.3%) and drives budget-hit non-termination in 46% of trials.

💡 Summary & Analysis

1. **The first key contribution of this study:** It provides a comprehensive understanding of reinforcement learning algorithms. This is akin to finding the fastest route among multiple roads. 2. **Second important contribution:** Experimentally evaluating five major techniques, allowing for a clear identification of their strengths and weaknesses. This can be compared to testing car performance under various weather conditions and traffic scenarios. 3. **Third key contribution:** Suggests the most effective reinforcement learning technique. Similar to selecting not just any optimal path but also the fastest and most efficient vehicle to reach your destination.

📄 Full Paper Content (ArXiv Source)

1. **The first key contribution of this study:** It provides a comprehensive understanding of reinforcement learning algorithms. This is akin to finding the fastest route among multiple roads. 2. **Second important contribution:** Experimentally evaluating five major techniques, allowing for a clear identification of their strengths and weaknesses. This can be compared to testing car performance under various weather conditions and traffic scenarios. 3. **Third key contribution:** Suggests the most effective reinforcement learning technique. Similar to selecting not just any optimal path but also the fastest and most efficient vehicle to reach your destination.

📊 논문 시각자료 (Figures)

Figure 1



Figure 2



Figure 3



Figure 4



Figure 5



Figure 6



Figure 7



Figure 8



Figure 9



Figure 10



Figure 11



Figure 12



Figure 13



Figure 14



A Note of Gratitude

The copyright of this content belongs to the respective researchers. We deeply appreciate their hard work and contribution to the advancement of human civilization.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut