M2MeT challenge baseline -- AliMeeting
This project provides the baseline system recipes for the ICASSP 2020 Multi-channel Multi-party Meeting Transcription Challenge (M2MeT). The challenge mainly consists of two tracks, named Automatic Speech Recognition (ASR) and Speaker Diarization. For each track, detailed descriptions can be found in its corresponding directory. The goal of this project is to simplify the training and evaluation procedures and make it flexible for participants to reproduce the baseline experiments and develop novelty methods.
Setup
git clone https://github.com/yufan-aslp/AliMeeting.git
Introduction
- Speech Recognition Track: Follow the detailed steps in
./asr
. - Speaker Diarization Track: Follow the detailed steps in
./speaker
.
General steps
- Prepare the training data for speaker diarization and ASR model, respectively
- Follow the running steps of the speaker diarization experiment and obtain the
rttm
file. Therttm
file includes the voice activity detection (VAD) and speaker diarization results, which will be used to compute the final Diarization Error Rate (DER) scores. - For ASR track, we can train the single-speaker or multi-speaker ASR models. The evaluation metric of ASR systems is Character Error Rate (CER).
Citation
If you use the challenge dataset or our baseline systems, please consider citing the following:
@article{yu2021m2met,
title={M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge},
author={Yu, Fan and Zhang, Shiliang and Fu, Yihui and Xie, Lei and Zheng, Siqi and Du, Zhihao and Huang, Weilong and Guo, Pengcheng and Yan, Zhijie and Ma, Bin and others},
journal={arXiv preprint arXiv:2110.07393},
year={2021}
}
Our paper is available at https://arxiv.org/abs/2110.07393
The data download method will be sent to registered challenge participants via email.
Organizing Committee
- Lei Xie, AISHELL Foundation, China, [email protected]
- Bin Ma, Principal Engineer at Alibaba, Singapore, [email protected]
- DeLiang Wang, Professor, Ohio State University, USA, [email protected]
- Zheng-Hua Tan, Professor, Aalborg University, Denmark, [email protected]
- Kong Aik Lee, Senior Scientist, Institute for Infocomm Research, A*STAR, Singapore, [email protected]
- Zhijie Yan, Director of Speech Lab at Alibaba, China, [email protected]
- Yanmin Qian, Associate Professor, Shanghai Jiao Tong University, China, [email protected]
- Hui Bu, CEO, AIShell Inc., China, [email protected]