When, How Long and How Much? Interpretable Neural Networks for Time Series Regression by Learning to Mask and Aggregate

Reading time: 6 minute
...

📝 Original Info

  • Title: When, How Long and How Much? Interpretable Neural Networks for Time Series Regression by Learning to Mask and Aggregate
  • ArXiv ID: 2512.03578
  • Date: 2025-12-03
  • Authors: ** Florent Forest, Amaury Wei, Olga Fink **

📝 Abstract

Time series extrinsic regression (TSER) refers to the task of predicting a continuous target variable from an input time series. It appears in many domains, including healthcare, finance, environmental monitoring, and engineering. In these settings, accurate predictions and trustworthy reasoning are both essential. Although state-of-the-art TSER models achieve strong predictive performance, they typically operate as black boxes, making it difficult to understand which temporal patterns drive their decisions. Post-hoc interpretability techniques, such as feature attribution, aim to to explain how the model arrives at its predictions, but often produce coarse, noisy, or unstable explanations. Recently, inherently interpretable approaches based on concepts, additive decompositions, or symbolic regression, have emerged as promising alternatives. However, these approaches remain limited: they require explicit supervision on the concepts themselves, often cannot capture interactions between time-series features, lack expressiveness for complex temporal patterns, and struggle to scale to high-dimensional multivariate data. To address these limitations, we propose MAGNETS (Mask-and-AGgregate NEtwork for Time Series), an inherently interpretable neural architecture for TSER. MAGNETS learns a compact set of human-understandable concepts without requiring any annotations. Each concept corresponds to a learned, mask-based aggregation over selected input features, explicitly revealing both which features drive predictions and when they matter in the sequence. Predictions are formed as combinations of these learned concepts through a transparent, additive structure, enabling clear insight into the model's decision process. The code implementation and datasets are publicly available at https://github.com/FlorentF9/MAGNETS.

💡 Deep Analysis

Deep Dive into When, How Long and How Much? Interpretable Neural Networks for Time Series Regression by Learning to Mask and Aggregate.

Time series extrinsic regression (TSER) refers to the task of predicting a continuous target variable from an input time series. It appears in many domains, including healthcare, finance, environmental monitoring, and engineering. In these settings, accurate predictions and trustworthy reasoning are both essential. Although state-of-the-art TSER models achieve strong predictive performance, they typically operate as black boxes, making it difficult to understand which temporal patterns drive their decisions. Post-hoc interpretability techniques, such as feature attribution, aim to to explain how the model arrives at its predictions, but often produce coarse, noisy, or unstable explanations. Recently, inherently interpretable approaches based on concepts, additive decompositions, or symbolic regression, have emerged as promising alternatives. However, these approaches remain limited: they require explicit supervision on the concepts themselves, often cannot capture interactions between

📄 Full Content

1 When, How Long and How Much? Interpretable Neural Networks for Time Series Regression by Learning to Mask and Aggregate Florent Forest, Amaury Wei, Olga Fink Abstract—Time series extrinsic regression (TSER) refers to the task of predicting a continuous target variable from an input time series. It appears in many domains, including healthcare, finance, environmental monitoring, and engineering. In these settings, accurate predictions and trustworthy reasoning are both essential. Although state-of-the-art TSER models achieve strong predictive performance, they typically operate as black boxes, making it difficult to understand which temporal patterns drive their decisions. Post-hoc interpretability techniques, such as feature attribution, aim to to explain how the model arrives at its predictions, but often produce coarse, noisy, or unstable expla- nations. Recently, inherently interpretable approaches based on concepts, additive decompositions, or symbolic regression, have emerged as promising alternatives. However, these approaches remain limited: they require explicit supervision on the concepts themselves, often cannot capture interactions between time-series features, lack expressiveness for complex temporal patterns, and struggle to scale to high-dimensional multivariate data. To address these limitations, we propose MAGNETS (Mask- and-AGgregate NEtwork for Time Series), an inherently inter- pretable neural architecture for TSER. MAGNETS learns a compact set of human-understandable concepts without requir- ing any annotations. Each concept corresponds to a learned, mask-based aggregation over selected input features, explicitly revealing both which features drive predictions and when they matter in the sequence. Predictions are formed as combinations of these learned concepts through a transparent, additive structure, enabling clear insight into the model’s decision process. Experiments on synthetic and real-world univariate and mul- tivariate TSER datasets show that MAGNETS closely matches the accuracy of black-box models while substantially outperform- ing existing interpretable baselines, particularly on multivariate datasets involving feature interactions. Finally, we also show that MAGNETS provides more faithful and informative explanations than post-hoc methods. The code implementation and datasets are publicly available at https://github.com/FlorentF9/MAGNETS. Index Terms—Time series regression, Machine learning, Ex- plainability, Concept learning, Interpretability. I. INTRODUCTION T IME series extrinsic regression (TSER) refers to to the task of predicting a continuous target variable from an input time series [1]. It appears in many domains—including healthcare (e.g., vital-sign forecasting), finance (e.g., volatil- ity prediction), engineering (e.g., predicting battery state- of-charge or estimating remaining useful life from sensor streams), and environmental monitoring (e.g., pollution esti- mation). In these settings, accurate predictions and trustworthy reasoning are both essential [2]. However, despite recent gains in accuracy, the opacity of modern TSER models remains an obstacle to adoption, particularly when predictions must be understood and validated by domain experts. The demand for interpretability becomes especially difficult to satisfy with current state-of-the-art TSER models. Although recent approaches achieve strong predictive performance, they typically operate as black boxes. Leading approaches include deep neural networks [3], [4], large ensembles [5], and Ran- dOm Convolutional KErnel Transform (ROCKET) models [6]–[8]. Their complexity, large parameter counts, and lack of transparency make it difficult to understand which temporal features drive predictions or how multivariate interactions across variables influence the model’s output. These limita- tions hinder deployment in sensitive or regulated contexts. For example, if a TSER model predicts an impending system failure, engineers need to trace that prediction back to specific sensor behaviors and time intervals, rather than relying on an alert that cannot be meaningfully explained or validated. These interpretability challenges have motivated growing interest in Explainable AI (XAI) research for time series, which aims to bridge the gap between predictive accuracy and actionable insight. While this research direction has made meaningful progress, existing approaches still fall short in several ways. Post-hoc explanation methods, such as saliency maps and feature attribution [9]–[11], attempt to rationalize a model’s decisions after training, but their explanations are often coarse, unstable, or poorly aligned with the model’s true internal reasoning [12]. In contrast, inherently interpretable models embed trans- parency directly into their architecture. Neural additive models (NAMs) [13] and their time-series extension, Neural Addi- tive Time-series Models (NATMs) [14], offer interpretable decompositions but remain

…(Full text truncated)…

📸 Image Gallery

StaticBridge-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.png StaticBridge-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.webp StaticBridge-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx3.png StaticBridge-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx3.webp StaticBridge-cnn-lr3-bs8-cosine-attributions-dl-idx3.png StaticBridge-cnn-lr3-bs8-cosine-attributions-dl-idx3.webp amaury_wei.png amaury_wei.webp area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-cosine-bottleneck.png area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-cosine-bottleneck.webp area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.png area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.webp area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx1.png area_over_threshold-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx1.webp area_over_threshold-cnn-lr3-bs8-cosine-attributions-dl-idx1.png area_over_threshold-cnn-lr3-bs8-cosine-attributions-dl-idx1.webp area_over_threshold_multivariate-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.png area_over_threshold_multivariate-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-bottleneck.webp area_over_threshold_multivariate-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx1.png area_over_threshold_multivariate-cdm_learned_nc3-masks10-lr3-bs8-unet-onlysum-spars1.0-cortho1.0-cosine-overlay-masks-idx1.webp area_over_threshold_multivariate-cnn-lr3-bs8-cosine-attributions-dl-idx1.png area_over_threshold_multivariate-cnn-lr3-bs8-cosine-attributions-dl-idx1.webp bivariate-s.png bivariate-s.webp florent_forest.png florent_forest.webp magnets-overview.png magnets-overview.webp olga_fink.jpg olga_fink.webp trivariate-1-s.png trivariate-1-s.webp trivariate-2-s.png trivariate-2-s.webp univariate-s.png univariate-s.webp

Reference

This content is AI-processed based on ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut