When Attention Fails: Pitfalls of Attention-based Model Interpretability for High-dimensional Clinical Time-Series

Research output: Contribution to journalConference articlepeer-review

1 Scopus citations

Abstract

Attention-based deep learning models are widely used for clinical time-series analysis, largely due to their perceived ability to enhance model interpretability. However, the reliability and consistency of attention mechanisms as an interpretability tool in high-dimensional clinical time series data require further investigation. We conducted a comprehensive evaluation of consistency of attention mechanisms in deep learning models applied to highdimensional clinical time-series data. Specifically, we trained 1000 different variants1 of an attention-based LSTM model architecture with random initializations to analyze the consistency of attention scores across mortality prediction and patient severity group classification. Our findings revealed significant inconsistencies in attention scores for individual samples across the thousand model variants. Visual inspection of attention weight distributions indicated that the attention mechanism did not consistently focus on the same feature-time pairs, challenging the assumption of reliability in model interpretability. The observed inconsistencies in persample attention weights suggest that attention mechanisms are unreliable as an interpretability tool for clinical decision-making tasks involving high-dimensional time-series data. While attention mechanisms may enhance model performance metrics, they often fail to produce clinically meaningful and consistent interpretations, limiting their utility in healthcare settings where transparency is critical for informed decision-making.

Original languageEnglish (US)
JournalProceedings of Machine Learning Research
Volume287
StatePublished - 2025
Event6th Conference on Health, Inference, and Learning, CHIL 2025 - Berkeley, United States
Duration: Jun 25 2025Jun 27 2025

ASJC Scopus subject areas

  • Software
  • Control and Systems Engineering
  • Statistics and Probability
  • Artificial Intelligence

Fingerprint

Dive into the research topics of 'When Attention Fails: Pitfalls of Attention-based Model Interpretability for High-dimensional Clinical Time-Series'. Together they form a unique fingerprint.

Cite this