| | |
| |
|
| | import csv |
| | import os |
| | import yaml |
| | from itertools import groupby |
| | from pathlib import Path |
| |
|
| | import torchaudio |
| |
|
| | import datasets |
| |
|
| |
|
| | _VERSION = "3.0.0" |
| |
|
| | _CITATION = """ |
| | @article{CATTONI2021101155, |
| | title = {MuST-C: A multilingual corpus for end-to-end speech translation}, |
| | author = {Roldano Cattoni and Mattia Antonino {Di Gangi} and Luisa Bentivogli and Matteo Negri and Marco Turchi}, |
| | journal = {Computer Speech & Language}, |
| | volume = {66}, |
| | pages = {101155}, |
| | year = {2021}, |
| | issn = {0885-2308}, |
| | doi = {https://doi.org/10.1016/j.csl.2020.101155}, |
| | url = {https://www.sciencedirect.com/science/article/pii/S0885230820300887}, |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """ |
| | MuST-C is a multilingual speech translation corpus whose size and quality facilitates |
| | the training of end-to-end systems for speech translation from English into several languages. |
| | For each target language, MuST-C comprises several hundred hours of audio recordings |
| | from English [TED Talks](https://www.ted.com/talks), which are automatically aligned |
| | at the sentence level with their manual transcriptions and translations. |
| | """ |
| |
|
| | _HOMEPAGE = "https://ict.fbk.eu/must-c/" |
| |
|
| | _LANGUAGES = ["de", "ja", "zh"] |
| |
|
| | _SAMPLE_RATE = 16_000 |
| |
|
| |
|
| | class MUSTC(datasets.GeneratorBasedBuilder): |
| | """MUSTC Dataset.""" |
| |
|
| | VERSION = datasets.Version(_VERSION) |
| |
|
| | BUILDER_CONFIGS = [ |
| | datasets.BuilderConfig(name=f"en-{lang}", version=datasets.Version(_VERSION)) for lang in _LANGUAGES |
| | ] |
| |
|
| | @property |
| | def manual_download_instructions(self): |
| | return f"""Please download the MUST-C v3 from https://ict.fbk.eu/must-c/ |
| | and unpack it with `tar xvzf MUSTC_v3.0_{self.config.name}.tar.gz`. |
| | Make sure to pass the path to the directory in which you unpacked the downloaded |
| | file as `data_dir`: `datasets.load_dataset('mustc', data_dir="path/to/dir")` |
| | """ |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | client_id=datasets.Value("string"), |
| | file=datasets.Value("string"), |
| | audio=datasets.Audio(sampling_rate=_SAMPLE_RATE), |
| | sentence=datasets.Value("string"), |
| | translation=datasets.Value("string"), |
| | id=datasets.Value("string"), |
| | ), |
| | supervised_keys=("file", "translation"), |
| | homepage=_HOMEPAGE, |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | source_lang, target_lang = self.config.name.split("-") |
| | assert source_lang == "en" |
| | assert target_lang in _LANGUAGES |
| |
|
| | data_root = os.path.abspath(os.path.expanduser(dl_manager.manual_dir)) |
| | root_path = Path(data_root) / self.config.name |
| |
|
| | if not os.path.exists(root_path): |
| | raise FileNotFoundError( |
| | "Dataset not found. Manual download required. " |
| | f"{self.manual_download_instructions}" |
| | ) |
| |
|
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={"root_path": root_path, "split": "train"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={"root_path": root_path, "split": "dev"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split("tst.COMMON"), |
| | gen_kwargs={"root_path": root_path, "split": "tst-COMMON"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split("tst.HE"), |
| | gen_kwargs={"root_path": root_path, "split": "tst-HE"}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, root_path, split): |
| | source_lang, target_lang = self.config.name.split("-") |
| |
|
| | |
| | txt_root = Path(root_path) / "data" / split / "txt" |
| | with (txt_root / f"{split}.yaml").open("r") as f: |
| | segments = yaml.load(f, Loader=yaml.BaseLoader) |
| |
|
| | |
| | with open(txt_root / f"{split}.{source_lang}", "r") as s_f: |
| | with open(txt_root / f"{split}.{target_lang}", "r") as t_f: |
| | s_lines = s_f.readlines() |
| | t_lines = t_f.readlines() |
| | assert len(s_lines) == len(t_lines) == len(segments) |
| | for i, (src, trg) in enumerate(zip(s_lines, t_lines)): |
| | segments[i][source_lang] = src.rstrip() |
| | segments[i][target_lang] = trg.rstrip() |
| |
|
| | |
| | _id = 0 |
| | wav_root = Path(root_path) / "data" / split / "wav" |
| | for wav_filename, _seg_group in groupby(segments, lambda x: x["wav"]): |
| | wav_path = wav_root / wav_filename |
| | seg_group = sorted(_seg_group, key=lambda x: float(x["offset"])) |
| | for i, segment in enumerate(seg_group): |
| | offset = int(float(segment["offset"]) * int(_SAMPLE_RATE)) |
| | duration = int(float(segment["duration"]) * int(_SAMPLE_RATE)) |
| | waveform, sr = torchaudio.load(wav_path, |
| | frame_offset=offset, |
| | num_frames=duration) |
| | assert duration == waveform.size(1), (duration, waveform.size(1)) |
| | assert sr == int(_SAMPLE_RATE), (sr, int(_SAMPLE_RATE)) |
| |
|
| | yield _id, { |
| | "file": wav_path.as_posix(), |
| | "audio": { |
| | "array": waveform.squeeze().numpy(), |
| | "path": wav_path.as_posix(), |
| | "sampling_rate": sr, |
| | }, |
| | "sentence": segment[source_lang], |
| | "translation": segment[target_lang], |
| | "client_id": segment["speaker_id"], |
| | "id": f"{wav_path.stem}_{i}", |
| | } |
| | _id += 1 |
| |
|