🤖 AI Summary
Existing evaluation metrics for film dubbing fail to comprehensively capture multidimensional complexities—including dialogue, narration, monologue, and actor-voice alignment—and lack an industrial-grade, systematic benchmark. To address this, we propose TA-Dubbing, the first adaptive dubbing evaluation benchmark tailored for professional film production. It introduces a novel dual-path assessment framework that jointly models cinematic semantic understanding and speech generation quality, enabling multimodal, multi-character, and multi-context adaptive dubbing evaluation. The benchmark comprises an open-source video–text–speech aligned dataset, an extensible evaluation toolchain, and a dynamic leaderboard. Extensive experiments demonstrate its strong discriminative capability across state-of-the-art dubbing models and multimodal foundation models. TA-Dubbing advances standardized, industrially viable quality assessment for film dubbing.
📝 Abstract
Movie dubbing has advanced significantly, yet assessing the real-world effectiveness of these models remains challenging. A comprehensive evaluation benchmark is crucial for two key reasons: 1) Existing metrics fail to fully capture the complexities of dialogue, narration, monologue, and actor adaptability in movie dubbing. 2) A practical evaluation system should offer valuable insights to improve movie dubbing quality and advancement in film production. To this end, we introduce Talking Adaptive Dubbing Benchmarks (TA-Dubbing), designed to improve film production by adapting to dialogue, narration, monologue, and actors in movie dubbing. TA-Dubbing offers several key advantages: 1) Comprehensive Dimensions: TA-Dubbing covers a variety of dimensions of movie dubbing, incorporating metric evaluations for both movie understanding and speech generation. 2) Versatile Benchmarking: TA-Dubbing is designed to evaluate state-of-the-art movie dubbing models and advanced multi-modal large language models. 3) Full Open-Sourcing: We fully open-source TA-Dubbing at https://github.com/woka- 0a/DeepDubber- V1 including all video suits, evaluation methods, annotations. We also continuously integrate new movie dubbing models into the TA-Dubbing leaderboard at https://github.com/woka- 0a/DeepDubber-V1 to drive forward the field of movie dubbing.