Frequently asked questions will be listed below.
Each team can participate in only one track — either TTS or SASV. Participants cannot join both tracks at the same time.
For the SASV track, the test dataset will be released after November 15th, once we receive outputs from the TTS track teams.
Yes, there is a typo in Table 1 of the paper. The numbers of samples for TITW-Easy and TITW-Hard are reversed. The correct statistics are:
| Set | samples | Avg dur (s) | Tot dur (h) | Avg # words |
|---|---|---|---|---|
| TITW-Easy | 248,024 | 2.51 | 173 | 10.55 |
| TITW-Hard | 282,606 | 2.42 | 189 | 10.84 |
Yes, pretrained datasets and models are allowed. However, the final submitted model must be fine-tuned on the TITW dataset in the last phase.
Yes, combining multiple models during inference is allowed.
We will use Versa as our evaluation tool. WER, UTMOS, DNSMOS, and speaker similarity will be calculated using this tool automatically.
Only the VoxCeleb2 dataset can be used, as it does not overlap with the SpoofCeleb test set. No other external datasets are permitted for the SASV track.
Yes, SSL models are allowed.
Yes, SSL models pre-trained on other public datasets are permitted.
No, any model that applies self-supervised learning (SSL) is allowed, including but not limited to wav2vec 2.0, HuBERT, and other SSL-based models.
Yes, the use of multimodal foundation models, including LLMs, is allowed.
Yes, the use of MUSAN and RIR is permitted for data augmentation.
Yes, it’s acceptable as long as no additional external datasets are introduced.
Each test utterance will be paired with one registration utterance, following the provided evaluation protocol format (a list of <enroll test_trial > pairs).
A link has been sent to all registered teams. Full evaluation data will be made public after the challenge closes. If you haven’t received the link, please check the trash folder, or contact Yihan Wu (yihanwu@ruc.edu.cn).
If you have any other questions, please email the organizing committee: Yihan Wu (yihanwu@ruc.edu.cn).