Cross-Attention is Half Explanation in Speech-to-Text Models
Abstract
Cross-attention in speech-to-text models aligns moderately with saliency-based explanations but captures only a portion of input relevance and decoder attention.
Cross-attention is a core mechanism in encoder-decoder architectures, widespread in many fields, including speech-to-text (S2T) processing. Its scores have been repurposed for various downstream applications--such as timestamp estimation and audio-text alignment--under the assumption that they reflect the dependencies between input speech representation and the generated text. While the explanatory nature of attention mechanisms has been widely debated in the broader NLP literature, this assumption remains largely unexplored within the speech domain. To address this gap, we assess the explanatory power of cross-attention in S2T models by comparing its scores to input saliency maps derived from feature attribution. Our analysis spans monolingual and multilingual, single-task and multi-task models at multiple scales, and shows that attention scores moderately to strongly align with saliency-based explanations, particularly when aggregated across heads and layers. However, it also shows that cross-attention captures only about 50% of the input relevance and, in the best case, only partially reflects how the decoder attends to the encoder's representations--accounting for just 52-75% of the saliency. These findings uncover fundamental limitations in interpreting cross-attention as an explanatory proxy, suggesting that it offers an informative yet incomplete view of the factors driving predictions in S2T models.
Community
This paper investigates whether cross-attention in speech-to-text models can serve as a faithful explanation of model behavior. Through large-scale analysis against state-of-the-art saliency maps, we show that cross-attention captures only part of the explanation, cautioning against its sole use for interpretability and motivating the integration of complementary explanation methods.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Beyond Transcription: Mechanistic Interpretability in ASR (2025)
- DualSpeechLM: Towards Unified Speech Understanding and Generation via Dual Speech Token Modeling with Large Language Models (2025)
- MELA-TTS: Joint transformer-diffusion model with representation alignment for speech synthesis (2025)
- Whisper-UT: A Unified Translation Framework for Speech and Text (2025)
- Utilizing Multilingual Encoders to Improve Large Language Models for Low-Resource Languages (2025)
- Self-Improvement for Audio Large Language Model using Unlabeled Speech (2025)
- Attention2Probability: Attention-Driven Terminology Probability Estimation for Robust Speech-to-Text System (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper