Context-Selective State Space Models: Feedback is All You Need
📝 Original Info
- Title: Context-Selective State Space Models: Feedback is All You Need
- ArXiv ID: 2510.14027
- Date: 2025-10-15
- Authors: ** 논문에 명시된 저자 정보가 제공되지 않았습니다. (보통 “Authors: …” 형태로 기재됩니다.) **
📝 Abstract
Transformers, powered by the attention mechanism, are the backbone of most foundation models, yet they suffer from quadratic complexity and difficulties in dealing with long-range dependencies in the input sequence. Recent work has shown that state space models (SSMs) provide a promising alternative. In this paper, we introduce the COFFEE (COntext From FEEdback) model, a novel time-varying SSM that incorporates state feedback to enable context-dependent selectivity, while still allowing for parallel implementation. This idea allows the model to regulate its dynamics based on the context described by the internal state, which embodies a compact representation of the input history. State feedback allows COFFEE to improve its ability to capture long-range dependencies: on the induction head task, it achieves near-perfect accuracy with two orders of magnitude fewer parameters and training sequences compared to S6 (the SSM of Mamba). On MNIST, COFFEE largely outperforms S6 within the same architecture, reaching 97% accuracy with only 3585 parameters. These results showcase the role of state feedback as a key mechanism for building scalable and efficient sequence models.💡 Deep Analysis
📄 Full Content
Reference
This content is AI-processed based on open access ArXiv data.