🤖 AI Summary
This work investigates cross-modal interference from irrelevant audio—such as silence, synthetic noise, and environmental sounds—on the text reasoning capabilities of large audio-language models (LALMs). Contrary to assumptions, silence is not a neutral input; it exerts destabilizing effects comparable to noise, exposing an unexpected robustness challenge for audio-augmented language processing. Using multi-text reasoning benchmarks, we systematically vary audio type, duration, loudness, and decoding temperature to evaluate leading LALMs. We propose and validate prompt engineering and self-consistency strategies as mitigation approaches. Results show that all models suffer significant performance degradation, with larger models exhibiting greater—but insufficient—resilience. Self-consistency improves output stability at the cost of increased computational overhead. This study provides the first systematic empirical evidence of latent fragility in audio–text modality coupling, offering critical insights and actionable pathways for enhancing LALM robustness.
📝 Abstract
Large audio-language models (LALMs) unify speech and text processing, but their robustness in noisy real-world settings remains underexplored. We investigate how irrelevant audio, such as silence, synthetic noise, and environmental sounds, affects text reasoning tasks where audio is unnecessary. Across three text-based benchmarks, we find that even non-informative audio reduces accuracy and increases prediction volatility; the severity of interference scales with longer durations, higher amplitudes, and elevated decoding temperatures. Silence, often assumed neutral, destabilizes outputs as strongly as synthetic noise. While larger models show greater resilience, vulnerabilities persist across all evaluated systems. We further test mitigation strategies and find that prompting shows limited effectiveness, whereas self-consistency improves stability at the cost of increased computation. Our results reveal cross-modal interference as a key robustness challenge and highlight the need for efficient fusion strategies that preserve reasoning performance in the presence of irrelevant inputs.