Sensitivity-Aware Mixed-Precision Quantization for ReRAM-based Computing-in-Memory

Reading time: 1 minute
...

📝 Original Info

  • Title: Sensitivity-Aware Mixed-Precision Quantization for ReRAM-based Computing-in-Memory
  • ArXiv ID: 2512.19445
  • Date: 2025-12-22
  • Authors: ** 제공된 정보에 저자 명단이 포함되어 있지 않습니다. **

📝 Abstract

Compute-In-Memory (CIM) systems, particularly those utilizing ReRAM and memristive technologies, offer a promising path toward energy-efficient neural network computation. However, conventional quantization and compression techniques often fail to fully optimize performance and efficiency in these architectures. In this work, we present a structured quantization method that combines sensitivity analysis with mixed-precision strategies to enhance weight storage and computational performance on ReRAM-based CIM systems. Our approach improves ReRAM Crossbar utilization, significantly reducing power consumption, latency, and computational load, while maintaining high accuracy. Experimental results show 86.33% accuracy at 70% compression, alongside a 40% reduction in power consumption, demonstrating the method's effectiveness for power-constrained applications.

💡 Deep Analysis

Figure 1

📄 Full Content

📸 Image Gallery

Models_comparison.png fig1.png fig2.png fig3_1.png fig3_2.png fig3_3.png fig4.png fig5.png fig6.png fig7.png

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut