LLM Driven Processes to Foster Explainable AI
Reading time: 1 minute
...
📝 Original Info
- Title: LLM Driven Processes to Foster Explainable AI
- ArXiv ID: 2511.07086
- Date: 2025-11-10
- Authors: ** (논문에 명시된 저자 정보가 제공되지 않았으므로, 저자명은 “미상”으로 표기합니다.) **
📝 Abstract
We present a modular, explainable LLM-agent pipeline for decision support that externalizes reasoning into auditable artifacts. The system instantiates three frameworks: Vester's Sensitivity Model (factor set, signed impact matrix, systemic roles, feedback loops); normal-form games (strategies, payoff matrix, equilibria); and sequential games (role-conditioned agents, tree construction, backward induction), with swappable modules at every step. LLM components (default: GPT-5) are paired with deterministic analyzers for equilibria and matrix-based role classification, yielding traceable intermediates rather than opaque outputs. In a real-world logistics case (100 runs), mean factor alignment with a human baseline was 55.5\% over 26 factors and 62.9\% on the transport-core subset; role agreement over matches was 57\%. An LLM judge using an eight-criterion rubric (max 100) scored runs on par with a reconstructed human baseline. Configurable LLM pipelines can thus mimic expert workflows with transparent, inspectable steps.💡 Deep Analysis
📄 Full Content
Reference
This content is AI-processed based on open access ArXiv data.