Real and synthetic Punjabi speech datasets for automatic speech recognition

Satwinder Singh, Feng Hou, Ruili Wang

Research output: Journal PublicationArticlepeer-review

1 Citation (Scopus)

Abstract

Automatic speech recognition (ASR) has been an active area of research. Training with large annotated datasets is the key to the development of robust ASR systems. However, most available datasets are focused on high-resource languages like English, leaving a significant gap for low-resource languages. Among these languages is Punjabi, despite its large number of speakers, Punjabi lacks high-quality annotated datasets for accurate speech recognition. To address this gap, we introduce three labeled Punjabi speech datasets: Punjabi Speech (real speech dataset) and Google-synth/CMU-synth (synthesized speech datasets). The Punjabi Speech dataset consists of read speech recordings captured in various environments, including both studio and open settings. In addition, the Google-synth dataset is synthesized using Google's Punjabi text-to-speech cloud services. Furthermore, the CMU-synth dataset is created using the Clustergen model available in the Festival speech synthesis system developed by CMU. These datasets aim to facilitate the development of accurate Punjabi speech recognition systems, bridging the resource gap for this important language.

Original languageEnglish
Article number109865
JournalData in Brief
Volume52
DOIs
Publication statusPublished - Feb 2024
Externally publishedYes

Keywords

  • Automatic speech recognition
  • low-resource languages
  • Punjabi language
  • Speech dataset

ASJC Scopus subject areas

  • General

Fingerprint

Dive into the research topics of 'Real and synthetic Punjabi speech datasets for automatic speech recognition'. Together they form a unique fingerprint.

Cite this