| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | """Passage, query, answers and answer classification with explanations.""" |
| |
|
| |
|
| | import json |
| |
|
| | import datasets |
| |
|
| |
|
| | _CITATION = """ |
| | @unpublished{eraser2019, |
| | title = {ERASER: A Benchmark to Evaluate Rationalized NLP Models}, |
| | author = {Jay DeYoung and Sarthak Jain and Nazneen Fatema Rajani and Eric Lehman and Caiming Xiong and Richard Socher and Byron C. Wallace} |
| | } |
| | @inproceedings{MultiRC2018, |
| | author = {Daniel Khashabi and Snigdha Chaturvedi and Michael Roth and Shyam Upadhyay and Dan Roth}, |
| | title = {Looking Beyond the Surface:A Challenge Set for Reading Comprehension over Multiple Sentences}, |
| | booktitle = {NAACL}, |
| | year = {2018} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """ |
| | Eraser Multi RC is a dataset for queries over multi-line passages, along with |
| | answers and a rationalte. Each example in this dataset has the following 5 parts |
| | 1. A Mutli-line Passage |
| | 2. A Query about the passage |
| | 3. An Answer to the query |
| | 4. A Classification as to whether the answer is right or wrong |
| | 5. An Explanation justifying the classification |
| | """ |
| |
|
| | _DOWNLOAD_URL = "http://www.eraserbenchmark.com/zipped/multirc.tar.gz" |
| |
|
| |
|
| | class EraserMultiRc(datasets.GeneratorBasedBuilder): |
| | """Multi Sentence Reasoning with Explanations (Eraser Benchmark).""" |
| |
|
| | VERSION = datasets.Version("0.1.1") |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "passage": datasets.Value("string"), |
| | "query_and_answer": datasets.Value("string"), |
| | "label": datasets.features.ClassLabel(names=["False", "True"]), |
| | "evidences": datasets.features.Sequence(datasets.Value("string")), |
| | } |
| | ), |
| | supervised_keys=None, |
| | homepage="https://cogcomp.seas.upenn.edu/multirc/", |
| | citation=_CITATION, |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | """Returns SplitGenerators.""" |
| |
|
| | archive = dl_manager.download(_DOWNLOAD_URL) |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | |
| | gen_kwargs={"files": dl_manager.iter_archive(archive), "split_file": "multirc/train.jsonl"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | |
| | gen_kwargs={"files": dl_manager.iter_archive(archive), "split_file": "multirc/val.jsonl"}, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | |
| | gen_kwargs={"files": dl_manager.iter_archive(archive), "split_file": "multirc/test.jsonl"}, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, files, split_file): |
| | """Yields examples.""" |
| |
|
| | multirc_dir = "multirc/docs" |
| | docs = {} |
| | for path, f in files: |
| | docs[path] = f.read().decode("utf-8") |
| | for line in docs[split_file].splitlines(): |
| | row = json.loads(line) |
| | evidences = [] |
| |
|
| | for evidence in row["evidences"][0]: |
| | docid = evidence["docid"] |
| | evidences.append(evidence["text"]) |
| |
|
| | passage_file = "/".join([multirc_dir, docid]) |
| | passage_text = docs[passage_file] |
| |
|
| | yield row["annotation_id"], { |
| | "passage": passage_text, |
| | "query_and_answer": row["query"], |
| | "label": row["classification"], |
| | "evidences": evidences, |
| | } |
| |
|