| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """ |
| EVI is a challenging spoken multilingual dataset with 5,506 dialogues in English, Polish, and French |
| that can be used for benchmarking and developing knowledge-based enrolment, identification, and identification |
| for spoken dialogue systems. |
| """ |
|
|
|
|
| import csv |
| from datetime import datetime |
| import json |
| import os |
| import warnings |
|
|
| import datasets |
|
|
| logger = datasets.logging.get_logger(__name__) |
|
|
|
|
| _CITATION = """\ |
| @inproceedings{Spithourakis2022evi, |
| author = {Georgios P. Spithourakis and Ivan Vuli\'{c} and Micha\l{} Lis and I\~{n}igo Casanueva and Pawe\l{} Budzianowski}, |
| title = {{EVI}: Multilingual Spoken Dialogue Tasks and Dataset for Knowledge-Based Enrolment, Verification, and Identification}, |
| year = {2022}, |
| note = {Data available at https://github.com/PolyAI-LDN/evi-paper}, |
| url = {https://arxiv.org/abs/2204.13496}, |
| booktitle = {Findings of NAACL (publication pending)} |
| } |
| """ |
|
|
| _ALL_CONFIGS = sorted([ |
| "en-GB", "fr-FR", "pl-PL" |
| ]) |
|
|
| _LANGS = sorted(["en", "fr", "pl"]) |
|
|
| _DESCRIPTION = """ |
| EVI is a challenging spoken multilingual dataset with 5,506 dialogues in English, Polish, and French |
| that can be used for benchmarking and developing knowledge-based enrolment, identification, and identification |
| for spoken dialogue systems. |
| """ |
|
|
| _LICENSE = "CC-BY-4.0" |
|
|
| _HOMEPAGE = "https://github.com/PolyAI-LDN/evi-paper" |
|
|
| _BASE_URL = "https://huggingface.co/datasets/PolyAI/evi/resolve/main/data" |
|
|
| _TEXTS_URL = { |
| lang: os.path.join(_BASE_URL, f"dialogues.{lang.split('-')[0]}.tsv") for lang in _LANGS |
| } |
|
|
| _RECORDS_URL = { |
| lang: os.path.join(_BASE_URL, f"records.{lang.split('-')[0]}.csv") for lang in _LANGS |
| } |
|
|
| _BROKEN_URL = { |
| "en": os.path.join(_BASE_URL, "broken_en.txt") |
| } |
|
|
| _AUDIO_DATA_URL = "https://poly-public-data.s3.eu-west-2.amazonaws.com/evi-paper/audios.zip" |
|
|
| _VERSION = datasets.Version("0.0.1", "") |
|
|
|
|
| class EviConfig(datasets.BuilderConfig): |
| """BuilderConfig for EVI""" |
|
|
| def __init__( |
| self, name, *args, **kwargs |
| ): |
| super().__init__(name=name, *args, **kwargs) |
| self.languages = _LANGS if name == "all" else [name.split("-")[0]] |
|
|
|
|
| class Evi(datasets.GeneratorBasedBuilder): |
|
|
| DEFAULT_WRITER_BATCH_SIZE = 512 |
| BUILDER_CONFIGS = [EviConfig(name) for name in _ALL_CONFIGS + ["all"]] |
|
|
| def _info(self): |
| features = datasets.Features( |
| { |
| "language": datasets.ClassLabel(names=_LANGS), |
| "audio": datasets.Audio(sampling_rate=8_000), |
| "asr_transcription": datasets.Value("string"), |
| "dialogue_id": datasets.Value("string"), |
| "speaker_id": datasets.Value("string"), |
| "turn_id": datasets.Value("int32"), |
| "target_profile_id": datasets.Value("string"), |
| "asr_nbest": datasets.Sequence(datasets.Value("string")), |
| "path": datasets.Value("string"), |
| "postcode": datasets.Value("string"), |
| "name": datasets.Value("string"), |
| "dob": datasets.Value("date64"), |
| "name_first": datasets.Value("string"), |
| "name_last": datasets.Value("string"), |
| "sex": datasets.ClassLabel(names=["F", "M"]), |
| "email": datasets.Value("string"), |
| } |
| ) |
|
|
| return datasets.DatasetInfo( |
| version=_VERSION, |
| description=_DESCRIPTION, |
| license=_LICENSE, |
| citation=_CITATION, |
| features=features, |
| homepage=_HOMEPAGE |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| langs = self.config.languages |
| lang2records_urls = { |
| lang: _RECORDS_URL[lang] for lang in langs |
| } |
| lang2text_urls = { |
| lang: _TEXTS_URL[lang] for lang in langs |
| } |
|
|
| records_paths = dl_manager.download_and_extract(lang2records_urls) |
| text_paths = dl_manager.download_and_extract(lang2text_urls) |
| audio_data_path = dl_manager.download_and_extract(_AUDIO_DATA_URL) |
|
|
| broken_path = dl_manager.download_and_extract(_BROKEN_URL["en"]) if "en" in langs else None |
|
|
| return [ |
| datasets.SplitGenerator( |
| name=datasets.Split.TEST, |
| gen_kwargs={ |
| "audio_data_path": audio_data_path, |
| "text_paths": text_paths, |
| "records_paths": records_paths, |
| "broken_path": broken_path |
| }, |
| ) |
| ] |
|
|
| def _generate_examples(self, audio_data_path, text_paths, records_paths, broken_path=None): |
| if broken_path: |
| with open(broken_path, encoding="utf-8") as f: |
| broken_samples = set([line.strip() for line in f]) |
| else: |
| broken_samples = None |
|
|
| for lang, text_path in text_paths.items(): |
|
|
| records_path = records_paths[lang] |
| records = dict() |
| with open(records_path, encoding="utf-8") as fin: |
| records_reader = csv.DictReader( |
| fin, delimiter=",", skipinitialspace=True |
| ) |
| for row in records_reader: |
| records[row["scenario_id"]] = row |
| records[row["scenario_id"]]["dob"] = datetime.strptime(row["dob"], "%Y-%m-%d") |
| _ = records[row["scenario_id"]].pop("scenario_id") |
|
|
| with open(text_path, encoding="utf-8") as fin: |
| texts_reader = csv.DictReader( |
| fin, delimiter="\t", skipinitialspace=True |
| ) |
| for dictrow in texts_reader: |
| dialogue_id = dictrow["dialogue_id"] |
| turn_id = dictrow["turn_num"] |
| file_path = os.path.join( |
| "audios", |
| lang, |
| dialogue_id, |
| f'{turn_id}.wav' |
| ) |
| full_path = os.path.join(audio_data_path, file_path) |
| if broken_samples and file_path in broken_samples: |
| warnings.warn(f"{full_path} is broken, skipping it.") |
| continue |
| if not os.path.isfile(full_path): |
| warnings.warn(f"{full_path} not found, skipping it.") |
| continue |
|
|
| target_profile_id = dictrow["scenario_id"] |
| if target_profile_id not in records: |
| warnings.warn( |
| f""" |
| Record with scenario_id {target_profile_id} not found, ignoring this dialogue. |
| Full dialogue info: {dictrow} |
| """ |
| ) |
| continue |
|
|
| yield file_path, { |
| "language": lang, |
| "audio": str(full_path), |
| "dialogue_id": dialogue_id, |
| "speaker_id": dictrow["speaker_id"], |
| "turn_id": turn_id, |
| "target_profile_id": target_profile_id, |
| "asr_transcription": dictrow["transcription"], |
| "asr_nbest": json.loads(dictrow["nbest"]), |
| "path": file_path, |
| **records[target_profile_id] |
| } |
|
|