| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """The text2log dataset""" |
|
|
|
|
| import csv |
|
|
| import datasets |
|
|
|
|
| _CITATION = """\ |
| @INPROCEEDINGS{9401852, author={Levkovskyi, Oleksii and Li, Wei}, booktitle={SoutheastCon 2021}, title={Generating Predicate Logic Expressions from Natural Language}, year={2021}, volume={}, number={}, pages={1-8}, doi={10.1109/SoutheastCon45413.2021.9401852}} |
| """ |
|
|
| _DESCRIPTION = """\ |
| The dataset contains about 100,000 simple English sentences selected and filtered from enTenTen15 and their translation into First Order Logic (FOL) Lambda Dependency-based Compositional Semantics using ccg2lambda. |
| """ |
|
|
| _HOMEPAGE = "https://github.com/alevkov/text2log" |
|
|
| _LICENSE = "none provided" |
|
|
|
|
| _URLS = { |
| "csv": "https://raw.githubusercontent.com/apergo-ai/text2log/main/dat/text2log_clean.csv", |
| "zip": "https://raw.githubusercontent.com/apergo-ai/text2log/main/dat/text2log_clean.zip", |
| } |
|
|
|
|
| class Text2log(datasets.GeneratorBasedBuilder): |
| """Simple English sentences and FOL representations using LDbCS""" |
|
|
| VERSION = datasets.Version("1.0.0") |
|
|
| def _info(self): |
|
|
| features = datasets.Features( |
| { |
| "sentence": datasets.Value("string"), |
| "fol_translation": datasets.Value("string"), |
| } |
| ) |
| return datasets.DatasetInfo( |
| description=_DESCRIPTION, |
| supervised_keys=None, |
| features=features, |
| homepage=_HOMEPAGE, |
| license=_LICENSE, |
| citation=_CITATION, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| """Returns SplitGenerators.""" |
| test_path = dl_manager.download_and_extract(_URLS["csv"]) |
| return [ |
| datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": test_path}), |
| ] |
|
|
| def _generate_examples(self, filepath): |
| """Generate text2log dataset examples.""" |
| with open(filepath, encoding="utf-8") as csv_file: |
| csv_reader = csv.reader( |
| csv_file, quotechar='"', delimiter=";", quoting=csv.QUOTE_ALL, skipinitialspace=True |
| ) |
| next(csv_reader) |
| for id_, row in enumerate(csv_reader): |
| yield id_, { |
| "sentence": str(row[0]), |
| "fol_translation": str(row[1]), |
| } |
|
|