论文标题
ICASSP 2021声学回声取消挑战:数据集,测试框架和结果
ICASSP 2021 Acoustic Echo Cancellation Challenge: Datasets, Testing Framework, and Results
论文作者
论文摘要
ICASSP 2021声学回声取消挑战旨在刺激声学回声取消(AEC)领域的研究,这是语音增强的重要组成部分,并且仍然是音频通信和会议系统中的主要问题。许多最近的AEC研究报告了在合成数据集上的良好性能,其中火车和测试样品来自相同的基础分布。但是,AEC的性能在真实录音中通常会大大降低。同样,在在现实环境中发现的背景噪声和混响的情况下,大多数传统的客观指标,例如回声回报损失增强(ERLE)和语音质量(PESQ)的感知评估(PESQ)与主观语音质量测试的相关性不佳。在这项挑战中,我们开源了两个大型数据集,以在单一谈话和双重谈话方案下培训AEC模型。这些数据集由来自2500多个真实的音频设备和人类在真实环境中的录音以及合成数据集组成。我们开源了两个大型测试集,并为研究人员开了一个在线主观测试框架,以快速测试其结果。这项挑战的优胜者将根据在所有不同的单一谈话和双重谈话场景中获得的平均平均意见分数(MOS)选择。
The ICASSP 2021 Acoustic Echo Cancellation Challenge is intended to stimulate research in the area of acoustic echo cancellation (AEC), which is an important part of speech enhancement and still a top issue in audio communication and conferencing systems. Many recent AEC studies report good performance on synthetic datasets where the train and test samples come from the same underlying distribution. However, the AEC performance often degrades significantly on real recordings. Also, most of the conventional objective metrics such as echo return loss enhancement (ERLE) and perceptual evaluation of speech quality (PESQ) do not correlate well with subjective speech quality tests in the presence of background noise and reverberation found in realistic environments. In this challenge, we open source two large datasets to train AEC models under both single talk and double talk scenarios. These datasets consist of recordings from more than 2,500 real audio devices and human speakers in real environments, as well as a synthetic dataset. We open source two large test sets, and we open source an online subjective test framework for researchers to quickly test their results. The winners of this challenge will be selected based on the average Mean Opinion Score (MOS) achieved across all different single talk and double talk scenarios.