| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality |
| | labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised |
| | and unsupervised training. Around 40,000 hours of transcribed audio is first collected from audiobooks, podcasts |
| | and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, |
| | sports, etc. A new forced alignment and segmentation pipeline is proposed to create sentence segments suitable |
| | for speech recognition training, and to filter out segments with low-quality transcription. For system training, |
| | GigaSpeech provides five subsets of different sizes, 10h, 250h, 1000h, 2500h, and 10000h. |
| | For our 10,000-hour XL training subset, we cap the word error rate at 4% during the filtering/validation stage, |
| | and for all our other smaller training subsets, we cap it at 0%. The DEV and TEST evaluation sets, on the other hand, |
| | are re-processed by professional human transcribers to ensure high transcription quality. |
| | """ |
| |
|
| | import csv |
| | import os |
| |
|
| | import datasets |
| |
|
| | _CITATION = """\ |
| | @article{DBLP:journals/corr/abs-2106-06909, |
| | author = {Guoguo Chen and |
| | Shuzhou Chai and |
| | Guanbo Wang and |
| | Jiayu Du and |
| | Wei{-}Qiang Zhang and |
| | Chao Weng and |
| | Dan Su and |
| | Daniel Povey and |
| | Jan Trmal and |
| | Junbo Zhang and |
| | Mingjie Jin and |
| | Sanjeev Khudanpur and |
| | Shinji Watanabe and |
| | Shuaijiang Zhao and |
| | Wei Zou and |
| | Xiangang Li and |
| | Xuchen Yao and |
| | Yongqing Wang and |
| | Yujun Wang and |
| | Zhao You and |
| | Zhiyong Yan}, |
| | title = {GigaSpeech: An Evolving, Multi-domain {ASR} Corpus with 10, 000 Hours |
| | of Transcribed Audio}, |
| | journal = {CoRR}, |
| | volume = {abs/2106.06909}, |
| | year = {2021}, |
| | url = {https://arxiv.org/abs/2106.06909}, |
| | eprinttype = {arXiv}, |
| | eprint = {2106.06909}, |
| | timestamp = {Wed, 29 Dec 2021 14:29:26 +0100}, |
| | biburl = {https://dblp.org/rec/journals/corr/abs-2106-06909.bib}, |
| | bibsource = {dblp computer science bibliography, https://dblp.org} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality |
| | labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised |
| | and unsupervised training. Around 40,000 hours of transcribed audio is first collected from audiobooks, podcasts |
| | and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, |
| | sports, etc. A new forced alignment and segmentation pipeline is proposed to create sentence segments suitable |
| | for speech recognition training, and to filter out segments with low-quality transcription. For system training, |
| | GigaSpeech provides five subsets of different sizes, 10h, 250h, 1000h, 2500h, and 10000h. |
| | For our 10,000-hour XL training subset, we cap the word error rate at 4% during the filtering/validation stage, |
| | and for all our other smaller training subsets, we cap it at 0%. The DEV and TEST evaluation sets, on the other hand, |
| | are re-processed by professional human transcribers to ensure high transcription quality. |
| | """ |
| |
|
| | _HOMEPAGE = "https://groups.inf.ed.ac.uk/ami/corpus/" |
| |
|
| | _LICENSE = "CC BY 4.0" |
| |
|
| | _TRAIN_SAMPLE_IDS = [ |
| | "EN2001a", |
| | "EN2001b", |
| | "EN2001d", |
| | "EN2001e", |
| | "EN2003a", |
| | "EN2004a", |
| | "EN2005a", |
| | "EN2006a", |
| | "EN2006b", |
| | "EN2009b", |
| | "EN2009c", |
| | "EN2009d", |
| | "ES2002a", |
| | "ES2002b", |
| | "ES2002c", |
| | "ES2002d", |
| | "ES2003a", |
| | "ES2003b", |
| | "ES2003c", |
| | "ES2003d", |
| | "ES2005a", |
| | "ES2005b", |
| | "ES2005c", |
| | "ES2005d", |
| | "ES2006a", |
| | "ES2006b", |
| | "ES2006c", |
| | "ES2006d", |
| | "ES2007a", |
| | "ES2007b", |
| | "ES2007c", |
| | "ES2007d", |
| | "ES2008a", |
| | "ES2008b", |
| | "ES2008c", |
| | "ES2008d", |
| | "ES2009a", |
| | "ES2009b", |
| | "ES2009c", |
| | "ES2009d", |
| | "ES2010a", |
| | "ES2010b", |
| | "ES2010c", |
| | "ES2010d", |
| | "ES2012a", |
| | "ES2012b", |
| | "ES2012c", |
| | "ES2012d", |
| | "ES2013a", |
| | "ES2013b", |
| | "ES2013c", |
| | "ES2013d", |
| | "ES2014a", |
| | "ES2014b", |
| | "ES2014c", |
| | "ES2014d", |
| | "ES2015a", |
| | "ES2015b", |
| | "ES2015c", |
| | "ES2015d", |
| | "ES2016a", |
| | "ES2016b", |
| | "ES2016c", |
| | "ES2016d", |
| | "IB4005", |
| | "IN1001", |
| | "IN1002", |
| | "IN1005", |
| | "IN1007", |
| | "IN1008", |
| | "IN1009", |
| | "IN1012", |
| | "IN1013", |
| | "IN1014", |
| | "IN1016", |
| | "IS1000a", |
| | "IS1000b", |
| | "IS1000c", |
| | "IS1000d", |
| | "IS1001a", |
| | "IS1001b", |
| | "IS1001c", |
| | "IS1001d", |
| | "IS1002b", |
| | "IS1002c", |
| | "IS1002d", |
| | "IS1003a", |
| | "IS1003b", |
| | "IS1003c", |
| | "IS1003d", |
| | "IS1004a", |
| | "IS1004b", |
| | "IS1004c", |
| | "IS1004d", |
| | "IS1005a", |
| | "IS1005b", |
| | "IS1005c", |
| | "IS1006a", |
| | "IS1006b", |
| | "IS1006c", |
| | "IS1006d", |
| | "IS1007a", |
| | "IS1007b", |
| | "IS1007c", |
| | "IS1007d", |
| | "TS3005a", |
| | "TS3005b", |
| | "TS3005c", |
| | "TS3005d", |
| | "TS3006a", |
| | "TS3006b", |
| | "TS3006c", |
| | "TS3006d", |
| | "TS3007a", |
| | "TS3007b", |
| | "TS3007c", |
| | "TS3007d", |
| | "TS3008a", |
| | "TS3008b", |
| | "TS3008c", |
| | "TS3008d", |
| | "TS3009a", |
| | "TS3009b", |
| | "TS3009c", |
| | "TS3009d", |
| | "TS3010a", |
| | "TS3010b", |
| | "TS3010c", |
| | "TS3010d", |
| | "TS3011a", |
| | "TS3011b", |
| | "TS3011c", |
| | "TS3011d", |
| | "TS3012a", |
| | "TS3012b", |
| | "TS3012c", |
| | "TS3012d", |
| | ] |
| |
|
| | _VALIDATION_SAMPLE_IDS = [ |
| | "ES2011a", |
| | "ES2011c", |
| | "IB4001", |
| | "IB4003", |
| | "IB4010", |
| | "IS1008a", |
| | "IS1008c", |
| | "TS3004a", |
| | "TS3004c", |
| | "ES2011b", |
| | "ES2011d", |
| | "IB4002", |
| | "IB4004", |
| | "IB4011", |
| | "IS1008b", |
| | "IS1008d", |
| | "TS3004b", |
| | "TS3004d", |
| | ] |
| |
|
| | _EVAL_SAMPLE_IDS = [ |
| | "EN2002a", |
| | "EN2002b", |
| | "EN2002c", |
| | "EN2002d", |
| | "ES2004a", |
| | "ES2004b", |
| | "ES2004c", |
| | "ES2004d", |
| | "IS1009a", |
| | "IS1009b", |
| | "IS1009c", |
| | "IS1009d", |
| | "TS3003a", |
| | "TS3003b", |
| | "TS3003c", |
| | "TS3003d", |
| | ] |
| |
|
| | _SUBSETS = ("ihm",) |
| |
|
| | _BASE_DATA_URL = "https://huggingface.co/datasets/speech-seq2seq/ami/resolve/main/" |
| |
|
| | _AUDIO_ARCHIVE_URL = _BASE_DATA_URL + "audio/{subset}/{split}/{_id}.tar.gz" |
| |
|
| | _ANNOTATIONS_ARCHIVE_URL = _BASE_DATA_URL + "annotations/{split}/text" |
| |
|
| | logger = datasets.utils.logging.get_logger(__name__) |
| |
|
| |
|
| | class AMIConfig(datasets.BuilderConfig): |
| | """BuilderConfig for AMI.""" |
| |
|
| | def __init__(self, name, *args, **kwargs): |
| | """BuilderConfig for AMI""" |
| | super().__init__(name=name, *args, **kwargs) |
| |
|
| |
|
| | class AMI(datasets.GeneratorBasedBuilder): |
| | """ |
| | GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality |
| | labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised |
| | and unsupervised training (this implementation contains only labelled data for now). |
| | Around 40,000 hours of transcribed audio is first collected from audiobooks, podcasts |
| | and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science, |
| | sports, etc. A new forced alignment and segmentation pipeline is proposed to create sentence segments suitable |
| | for speech recognition training, and to filter out segments with low-quality transcription. For system training, |
| | GigaSpeech provides five subsets of different sizes, 10h, 250h, 1000h, 2500h, and 10000h. |
| | For our 10,000-hour XL training subset, we cap the word error rate at 4% during the filtering/validation stage, |
| | and for all our other smaller training subsets, we cap it at 0%. The DEV and TEST evaluation sets, on the other hand, |
| | are re-processed by professional human transcribers to ensure high transcription quality. |
| | """ |
| |
|
| | VERSION = datasets.Version("1.0.0") |
| |
|
| | BUILDER_CONFIGS = [ |
| | AMIConfig(name=subset) for subset in _SUBSETS |
| | ] |
| |
|
| | DEFAULT_WRITER_BATCH_SIZE = 128 |
| |
|
| | def _info(self): |
| | features = datasets.Features( |
| | { |
| | "segment_id": datasets.Value("string"), |
| | "audio_id": datasets.Value("string"), |
| | "text": datasets.Value("string"), |
| | "audio": datasets.Audio(sampling_rate=16_000), |
| | "begin_time": datasets.Value("float32"), |
| | "end_time": datasets.Value("float32"), |
| | "microphone_id": datasets.Value("string"), |
| | "speaker_id": datasets.Value("string"), |
| | } |
| | ) |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=features, |
| | homepage=_HOMEPAGE, |
| | license=_LICENSE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | train_audio_files = {m: _AUDIO_ARCHIVE_URL.format(subset=self.config.name, split="train", _id=m) for m in _TRAIN_SAMPLE_IDS} |
| | dev_audio_files = {m: _AUDIO_ARCHIVE_URL.format(subset=self.config.name, split="dev", _id=m) for m in _VALIDATION_SAMPLE_IDS} |
| | eval_audio_files = {m: _AUDIO_ARCHIVE_URL.format(subset=self.config.name, split="eval", _id=m) for m in _EVAL_SAMPLE_IDS} |
| |
|
| | train_audio_archives = dl_manager.download_and_extract(train_audio_files) |
| | dev_audio_archives = dl_manager.download_and_extract(dev_audio_files) |
| | eval_audio_archives = dl_manager.download_and_extract(eval_audio_files) |
| |
|
| | train_annotation = dl_manager.download_and_extract(_ANNOTATIONS_ARCHIVE_URL.format(split="train")) |
| | dev_annotation = dl_manager.download_and_extract(_ANNOTATIONS_ARCHIVE_URL.format(split="dev")) |
| | eval_annotation = dl_manager.download_and_extract(_ANNOTATIONS_ARCHIVE_URL.format(split="eval")) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={"audio": train_audio_archives, "annotation": train_annotation, "split": "train"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={"audio": dev_audio_archives, "annotation": dev_annotation, "split": "dev"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={"audio": eval_audio_archives, "annotation": eval_annotation, "split": "eval"}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, audio, annotation, split): |
| | |
| | with open(annotation, "r", encoding="utf-8") as f: |
| | transcriptions = {} |
| | for line in f.readlines(): |
| | line_items = line.strip().split() |
| | _id = line_items[0] |
| | text = " ".join(line_items[1:]) |
| | _, segment_id, microphone_id, speaker_id, begin_time, end_time = _id.split("_") |
| |
|
| | transcriptions[_id] = { |
| | "audio_id": _id, |
| | "segment_id": segment_id, |
| | "text": text, |
| | "begin_time": int(begin_time) / 100, |
| | "end_time": int(end_time) / 100, |
| | "microphone_id": microphone_id, |
| | "speaker_id": speaker_id, |
| | } |
| |
|
| | for _audio_id, (transcription_id, result) in enumerate(transcriptions.items()): |
| | folder_id = result["segment_id"] |
| | file_name = "_".join([split, transcription_id.lower()]) + ".wav" |
| | audio_file = os.path.join(audio[folder_id], folder_id, file_name) |
| | result["audio"] = audio_file |
| | yield _audio_id, result |
| |
|
| |
|