Efficient Kernel Mapping and Comprehensive System Evaluation of LLM Acceleration on a CGLA

Reading time: 6 minute
...

📝 Original Info

  • Title: Efficient Kernel Mapping and Comprehensive System Evaluation of LLM Acceleration on a CGLA
  • ArXiv ID: 2512.00335
  • Date: 2025-11-29
  • Authors: Takuto Ando, Yu Eto, Ayumu Takeuchi, Yasuhiko Nakashima

📝 Abstract

Large Language Models (LLMs) demand substantial computational resources, resulting in high energy consumption on GPUs. To address this challenge, we focus on Coarse-Grained Reconfigurable Arrays (CGRAs) as an effective alternative that provides a trade-off between energy efficiency and programmability. This paper presents the first comprehensive, end-to-end evaluation of a non-AI-specialized Coarse-Grained Linear Array (CGLA) accelerator for the state-of-the-art Qwen3 LLM family. The architecture has a general-purpose, task-agnostic design, yet its flexible instruction set allows for domain-specific adaptations. This flexibility enables the architecture to achieve high efficiency for sustainable LLM inference. We assess the performance of our architecture on an FPGA prototype using the widely adopted llama.cpp framework. We then project its potential as a 28 nm ASIC and compare it against a high-performance GPU (NVIDIA RTX 4090) and an edge AI device (NVIDIA Jetson AGX Orin). While GPUs exhibit lower latency, our non-AI-specific accelerator achieves higher energy efficiency, improving the Power-Delay Product (PDP) by up to 44.4× and 13.6× compared with the RTX 4090 and Jetson, respectively. Similarly, it reduces the Energy-Delay Product (EDP) by up to 11.5× compared to the high-performance GPU, demonstrating a favorable performance-energy trade-off. Critically, our system-level analysis identifies host-accelerator data transfer as the primary performance bottleneck, a factor often overlooked in kernel-level studies. These findings provide design guidance for next-generation LLM accelerators. This work validates CGRAs as a suitable platform for LLM inference in power-constrained environments, without being confined to specific algorithms.

💡 Deep Analysis

Deep Dive into Efficient Kernel Mapping and Comprehensive System Evaluation of LLM Acceleration on a CGLA.

Large Language Models (LLMs) demand substantial computational resources, resulting in high energy consumption on GPUs. To address this challenge, we focus on Coarse-Grained Reconfigurable Arrays (CGRAs) as an effective alternative that provides a trade-off between energy efficiency and programmability. This paper presents the first comprehensive, end-to-end evaluation of a non-AI-specialized Coarse-Grained Linear Array (CGLA) accelerator for the state-of-the-art Qwen3 LLM family. The architecture has a general-purpose, task-agnostic design, yet its flexible instruction set allows for domain-specific adaptations. This flexibility enables the architecture to achieve high efficiency for sustainable LLM inference. We assess the performance of our architecture on an FPGA prototype using the widely adopted llama.cpp framework. We then project its potential as a 28 nm ASIC and compare it against a high-performance GPU (NVIDIA RTX 4090) and an edge AI device (NVIDIA Jetson AGX Orin). While GPU

📄 Full Content

Received 8 October 2025, accepted 19 November 2025. Digital Object Identifier 10.1109/ACCESS.2025.3636266 Efficient Kernel Mapping and Comprehensive System Evaluation of LLM Acceleration on a CGLA TAKUTO ANDO , (Member, IEEE), YU ETO , AYUMU TAKEUCHI and YASUHIKO NAKASHIMA , (Member, IEEE) Nara Institute of Science and Technology (NAIST), Ikoma, Nara 630-0192, Japan Corresponding author: Takuto ANDO (e-mail: antaku7585@gmail.com). This work was supported by the JST-ALCA-Next Program (Grant Number JPMJAN23F4) and JSPS KAKENHI (Grant No. 22H00515). We also acknowledge the activities of VDEC, The University of Tokyo, in collaboration with NIHON SYNOPSYS G.K. ABSTRACT Large Language Models (LLMs) demand substantial computational resources, resulting in high energy con- sumption on GPUs. To address this challenge, we focus on Coarse-Grained Reconfigurable Arrays (CGRAs) as an effective alternative that provides a trade-off between energy efficiency and programmability. This paper presents the first comprehensive, end-to-end evaluation of a non-AI-specialized Coarse-Grained Linear Array (CGLA) accelerator for the state-of-the-art Qwen3 LLM family. The architecture has a general-purpose, task-agnostic design, yet its flexible instruction set allows for domain-specific adaptations. This flexibility enables the architecture to achieve high efficiency for sustainable LLM inference. We assess the performance of our architecture on an FPGA prototype using the widely adopted llama.cpp framework. We then project its potential as a 28 nm ASIC and compare it against a high-performance GPU (NVIDIA RTX 4090) and an edge AI device (NVIDIA Jetson AGX Orin). While GPUs exhibit lower latency, our non-AI-specific accelerator achieves higher energy efficiency, improving the Power-Delay Product (PDP) by up to 44.4× and 13.6× compared with the RTX 4090 and Jetson, respectively. Similarly, it reduces the Energy-Delay Product (EDP) by up to 11.5× compared to the high-performance GPU, demonstrating a favorable performance-energy trade-off. Critically, our system-level analysis identifies host-accelerator data transfer as the primary performance bottleneck, a factor often overlooked in kernel-level studies. These findings provide design guidance for next-generation LLM accelerators. This work validates CGRAs as a suitable platform for LLM inference in power-constrained environments, without being confined to specific algorithms. INDEX TERMS Qwen, LLM, CGRA, CGLA, IMAX I. INTRODUCTION Large Language Models (LLMs) have evolved from spe- cialized text processing tools into a widely applied tech- nology in various fields [1], [2], [3], [4], [5]. They have been applied in fields such as natural language process- ing [6], [7], [8], code generation [9], [10], [11], and con- versational AI [12], [13]. The development of multimodal LLMs, which integrate modalities such as text, images, and audio, is further expanding their scope of application on a large scale [14], [15], [16]. The interaction between al- gorithmic development and computational architectures en- ables these innovations. Therefore, improving this infrastruc- ture is not merely a practical challenge but critical for the broader societal integration of AI. Currently, this technolog- ical progress primarily depends on the evolution of General- Purpose Graphics Processing Units (GPGPUs). However, despite these advances, achieving high performance with LLMs requires substantial computational resources for both training and inference. In particular, the high performance of GPGPUs requires substantial power, which creates sig- nificant environmental and economic concerns [17], [18]. The International Energy Agency (IEA) estimates that data center electricity consumption could double by 2030, reach- ing approximately 945 TWh [19]. This consumption level, VOLUME 13, 2025 1 arXiv:2512.00335v1 [cs.AR] 29 Nov 2025 T. Ando et al.: Efficient Kernel Mapping and Comprehensive System Evaluation of LLM Acceleration on a CGLA fueled primarily by the expanding demand for AI, slightly exceeds Japan’s total annual electricity consumption. This sustainability challenge presents a significant barrier to the widespread adoption of LLM technology, making it nec- essary to address this with improvements in computational architecture. To address this challenge, specialized hardware such as Application-Specific Integrated Circuits (ASICs) and Field- Programmable Gate Arrays (FPGAs) have been widely in- vestigated. While edge GPUs, such as the NVIDIA Jetson series, demonstrate progress in reducing power consumption, their general-purpose graphics pipelines inherently constrain substantial gains in power efficiency. In contrast, ASICs can potentially achieve orders of magnitude higher power efficiency. ASICs specialize their circuits for the core com- putational patterns of LLM inference, such as dot-product operations. This allows them to remove all unnecessary func- tionality and achieve maximum performanc

…(Full text truncated)…

📸 Image Gallery

E2E_latency_for_Qwen3-0_6B_-_Q3_K_S.png E2E_latency_for_Qwen3-0_6B_-_Q3_K_S.webp E2E_latency_for_Qwen3-0_6B_-_Q8_0.png E2E_latency_for_Qwen3-0_6B_-_Q8_0.webp E2E_latency_for_Qwen3-1_7B_-_Q3_K_S.png E2E_latency_for_Qwen3-1_7B_-_Q3_K_S.webp E2E_latency_for_Qwen3-1_7B_-_Q8_0.png E2E_latency_for_Qwen3-1_7B_-_Q8_0.webp E2E_latency_for_Qwen3-8B_-_Q3_K_S.png E2E_latency_for_Qwen3-8B_-_Q3_K_S.webp E2E_latency_for_Qwen3-8B_-_Q8_0.png E2E_latency_for_Qwen3-8B_-_Q8_0.webp EDP_for_Qwen3-0_6B_-_Q3_K_S.png EDP_for_Qwen3-0_6B_-_Q3_K_S.webp EDP_for_Qwen3-0_6B_-_Q8_0.png EDP_for_Qwen3-0_6B_-_Q8_0.webp EDP_for_Qwen3-1_7B_-_Q3_K_S.png EDP_for_Qwen3-1_7B_-_Q3_K_S.webp EDP_for_Qwen3-1_7B_-_Q8_0.png EDP_for_Qwen3-1_7B_-_Q8_0.webp EDP_for_Qwen3-8B_-_Q3_K_S.png EDP_for_Qwen3-8B_-_Q3_K_S.webp EDP_for_Qwen3-8B_-_Q8_0.png EDP_for_Qwen3-8B_-_Q8_0.webp PDP_for_Qwen3-0_6B_-_Q3_K_S.png PDP_for_Qwen3-0_6B_-_Q3_K_S.webp PDP_for_Qwen3-0_6B_-_Q8_0.png PDP_for_Qwen3-0_6B_-_Q8_0.webp PDP_for_Qwen3-1_7B_-_Q3_K_S.png PDP_for_Qwen3-1_7B_-_Q3_K_S.webp PDP_for_Qwen3-1_7B_-_Q8_0.png PDP_for_Qwen3-1_7B_-_Q8_0.webp PDP_for_Qwen3-8B_-_Q3_K_S.png PDP_for_Qwen3-8B_-_Q3_K_S.webp PDP_for_Qwen3-8B_-_Q8_0.png PDP_for_Qwen3-8B_-_Q8_0.webp PDP_vs_LMM_Qwen3-0.6B_Q3_K_S.png PDP_vs_LMM_Qwen3-0.6B_Q3_K_S.webp PDP_vs_LMM_Qwen3-0.6B_Q8_0.png PDP_vs_LMM_Qwen3-0.6B_Q8_0.webp PDP_vs_LMM_Qwen3-1.7B_Q3_K_S.png PDP_vs_LMM_Qwen3-1.7B_Q3_K_S.webp PDP_vs_LMM_Qwen3-1.7B_Q8_0.png PDP_vs_LMM_Qwen3-1.7B_Q8_0.webp PDP_vs_LMM_Qwen3-8B_Q3_K_S.png PDP_vs_LMM_Qwen3-8B_Q3_K_S.webp PDP_vs_LMM_Qwen3-8B_Q8_0.png PDP_vs_LMM_Qwen3-8B_Q8_0.webp ando2.png ando2.webp eto3.png eto3.webp f16dot.png f16dot.webp imax3_board_config.png imax3_board_config.webp internal_pe.png internal_pe.webp llama_arch.png llama_arch.webp logo.png logo.webp nakashima2.png nakashima2.webp notaglinelogo.png notaglinelogo.webp pe_detailed.png pe_detailed.webp performance_metrics_normalized.png performance_metrics_normalized.webp q3kdot.png q3kdot.webp q6kdot.png q6kdot.webp q8dot.png q8dot.webp q8dot_one_unit.png q8dot_one_unit.webp takeuchi.png takeuchi.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut