| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | """The Something-Something dataset (version 2) is a collection of 220,847 labeled video clips of humans performing pre-defined, basic actions with everyday objects.""" |
| |
|
| |
|
| | import csv |
| | import json |
| | import os |
| |
|
| | import datasets |
| |
|
| | from .classes import SOMETHING_SOMETHING_V2_CLASSES |
| |
|
| | _CITATION = """ |
| | @inproceedings{goyal2017something, |
| | title={The" something something" video database for learning and evaluating visual common sense}, |
| | author={Goyal, Raghav and Ebrahimi Kahou, Samira and Michalski, Vincent and Materzynska, Joanna and Westphal, Susanne and Kim, Heuna and Haenel, Valentin and Fruend, Ingo and Yianilos, Peter and Mueller-Freitag, Moritz and others}, |
| | booktitle={Proceedings of the IEEE international conference on computer vision}, |
| | pages={5842--5850}, |
| | year={2017} |
| | } |
| | """ |
| |
|
| | _DESCRIPTION = """\ |
| | The Something-Something dataset (version 2) is a collection of 220,847 labeled video clips of humans performing pre-defined, basic actions with everyday objects. It is designed to train machine learning models in fine-grained understanding of human hand gestures like putting something into something, turning something upside down and covering something with something. |
| | """ |
| |
|
| |
|
| | class SomethingSomethingV2(datasets.GeneratorBasedBuilder): |
| | """Charades is dataset composed of 9848 videos of daily indoors activities collected through Amazon Mechanical Turk""" |
| |
|
| | BUILDER_CONFIGS = [datasets.BuilderConfig(name="default")] |
| | DEFAULT_CONFIG_NAME = "default" |
| |
|
| | def _info(self): |
| | return datasets.DatasetInfo( |
| | description=_DESCRIPTION, |
| | features=datasets.Features( |
| | { |
| | "video_id": datasets.Value("string"), |
| | "video": datasets.Value("string"), |
| | "text": datasets.Value("string"), |
| | "label": datasets.features.ClassLabel( |
| | num_classes=len(SOMETHING_SOMETHING_V2_CLASSES), |
| | names=SOMETHING_SOMETHING_V2_CLASSES, |
| | ), |
| | "placeholders": datasets.Sequence(datasets.Value("string")), |
| | } |
| | ), |
| | supervised_keys=None, |
| | homepage="", |
| | citation=_CITATION, |
| | ) |
| |
|
| | @property |
| | def manual_download_instructions(self): |
| | return ( |
| | "To use Something-Something-v2, please download the 19 data files and the labels file " |
| | "from 'https://developer.qualcomm.com/software/ai-datasets/something-something'. " |
| | "Unzip the 19 files and concatenate the extracts in order into a tar file named '20bn-something-something-v2.tar.gz. " |
| | "Use command like `cat 20bn-something-something-v2-?? >> 20bn-something-something-v2.tar.gz` " |
| | "Place the `labels.zip` file and the tar file into a folder '/path/to/data/' and load the dataset using " |
| | "`load_dataset('something-something-v2', data_dir='/path/to/data')`" |
| | ) |
| |
|
| | def _split_generators(self, dl_manager): |
| | data_dir = dl_manager.manual_dir |
| | labels_path = os.path.join(data_dir, "labels.zip") |
| | videos_path = os.path.join(data_dir, "20bn-something-something-v2.tar.gz") |
| | if not os.path.exists(labels_path): |
| | raise FileNotFoundError( |
| | f"labels.zip doesn't exist in {data_dir}. Please follow manual download instructions." |
| | ) |
| |
|
| | if not os.path.exists(videos_path): |
| | raise FileNotFoundError( |
| | f"20bn-something-sokmething-v2.tar.gz doesn't exist in {data_dir}. Please follow manual download instructions." |
| | ) |
| |
|
| | labels_path = dl_manager.extract(labels_path) |
| | return [ |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TRAIN, |
| | gen_kwargs={ |
| | "annotation_file": os.path.join( |
| | labels_path, "labels", "train.json" |
| | ), |
| | "video_files": dl_manager.iter_archive(videos_path), |
| | }, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.VALIDATION, |
| | gen_kwargs={ |
| | "annotation_file": os.path.join( |
| | labels_path, "labels", "validation.json" |
| | ), |
| | "video_files": dl_manager.iter_archive(videos_path), |
| | }, |
| | ), |
| | datasets.SplitGenerator( |
| | name=datasets.Split.TEST, |
| | gen_kwargs={ |
| | "annotation_file": os.path.join(labels_path, "labels", "test.json"), |
| | "video_files": dl_manager.iter_archive(videos_path), |
| | "labels_file": os.path.join( |
| | labels_path, "labels", "test-answers.csv" |
| | ), |
| | }, |
| | ), |
| | ] |
| |
|
| | def _generate_examples(self, annotation_file, video_files, labels_file=None): |
| | data = {} |
| | labels = None |
| | if labels_file is not None: |
| | with open(labels_file, "r", encoding="utf-8") as fobj: |
| | labels = {} |
| | for label in fobj.readlines(): |
| | label = label.strip().split(";") |
| | labels[label[0]] = label[1] |
| |
|
| | with open(annotation_file, "r", encoding="utf-8") as fobj: |
| | annotations = json.load(fobj) |
| | for annotation in annotations: |
| | if "template" in annotation: |
| | annotation["template"] = ( |
| | annotation["template"].replace("[", "").replace("]", "") |
| | ) |
| | if labels: |
| | annotation["template"] = labels[annotation["id"]] |
| | data[annotation["id"]] = annotation |
| |
|
| | idx = 0 |
| | for path, file in video_files: |
| | video_id = os.path.splitext(os.path.split(path)[1])[0] |
| |
|
| | if video_id not in data: |
| | continue |
| |
|
| | info = data[video_id] |
| | yield idx, { |
| | "video_id": video_id, |
| | "video": file, |
| | "placeholders": info.get("placeholders", []), |
| | "label": info["label"] if "label" in info else -1, |
| | "text": info["template"], |
| | } |
| |
|
| | idx += 1 |
| |
|