Datasets
Standard Dataset
English as a Second Language TTS (ESLTTS) dataset
- Citation Author(s):
- Submitted by:
- WENBIN WANG
- Last updated:
- Sat, 04/27/2024 - 06:02
- DOI:
- 10.21227/cw1v-7p40
- Data Format:
- License:
- Categories:
- Keywords:
Abstract
With the progress made in speaker-adaptive TTS approaches, advanced approaches have shown a remarkable capacity to reproduce the speaker’s voice in the commonly used TTS datasets. However, mimicking voices characterized by substantial accents, such as non-native English speakers, is still challenging. Regrettably, the absence of a dedicated TTS dataset for speakers with substantial accents inhibits the research and evaluation of speaker-adaptive TTS models under such conditions. To address this gap, we developed a corpus of non-native speakers' English utterances.
We named this corpus “English as a Second Language TTS dataset ” (ESLTTS). The ESLTTS dataset consists of roughly 37 hours of 42,000 utterances from 134 non-native English speakers. These speakers represent a diversity of linguistic backgrounds spanning 31 native languages. For each speaker, the dataset includes an adaptation set lasting about 5 minutes for speaker adaptation, a test set comprising 10 utterances for speaker-adaptive TTS evaluation, and a development set for further research.
Dataset Structure:
ESLTTS Dataset/
├─ Malayalam_3/ ------------ {Speaker Native Language}_{Speaker id}
│ ├─ ada_1.flac ------------ {Subset Name}_{Utterance id}
│ ├─ ada_1.txt ------------ Transcription for "ada_1.flac"
│ ├─ test_1.flac ------------ {Subset Name}_{Utterance id}
│ ├─ test_1.txt ------------ Transcription for "test_1.flac"
│ ├─ dev_1.flac ------------ {Subset Name}_{Utterance id}
│ ├─ dev_1.txt ------------ Transcription for "dev_1.flac"
│ ├─ ...
├─ Arabic_3/ ------------ {Speaker Native Language}_{Speaker id}
│ ├─ ada_1.flac ------------ {Subset Name}_{Utterance id}
│ ├─ ...
├─ ...
Data Info:
Utterance Sample Rate: 24 kHz