dataset_name stringlengths 2 128 | description stringlengths 1 9.7k | prompt stringlengths 59 185 |
|---|---|---|
OSTD | This dataset consists of 18 movies with duration range between 10 and 104 minutes leveraged from the OVSD dataset (Rotman et al., 2016). For these videos, the summary length limit is set to be the minimum between 4 minutes and 10% of the video length. | Provide a detailed description of the following dataset: OSTD |
PolarRR | PolarRR is a new dataset with more than 100 types of glass in which obtained transmission images are perfectly aligned with input mixed images. | Provide a detailed description of the following dataset: PolarRR |
Lytro Illum | Lytro Illum is a new light field dataset using a Lytro Illum camera. 640 light fields are collected with significant variations in terms of size, textureness, background clutter and illumination, etc. Micro-lens image arrays and central viewing images are generated, and corresponding ground-truth maps are produced. | Provide a detailed description of the following dataset: Lytro Illum |
UFPR-Eyeglasses | The UFPR-Eyeglasses dataset has 1,135 images of both eyes (2,270 cropped images of each eye) from 83 subjects (166 classes). The dataset is used to evaluate the effect of the occlusion caused by eyeglasses in periocular recognition. | Provide a detailed description of the following dataset: UFPR-Eyeglasses |
Circa | The Circa (meaning ‘approximately’) dataset aims to help machine learning systems to solve the problem of interpreting indirect answers to polar questions.
The dataset contains pairs of yes/no questions and indirect answers, together with annotations for the interpretation of the answer. The data is collected in 10 ... | Provide a detailed description of the following dataset: Circa |
QUVA Repetition | QUVA Repetition dataset consists of 100 videos displaying a wide variety of repetitive video dynamics, including swimming, stirring, cutting, combing and music-making. All videos have been annotated with individual cycle bounds and a total repetition count. | Provide a detailed description of the following dataset: QUVA Repetition |
NH-HAZE | NN-HAZE is an image dehazing dataset. Since in many real cases haze is not uniformly distributed NH-HAZE, a non-homogeneous realistic dataset with pairs of real hazy and corresponding haze-free images. This is the first non-homogeneous image dehazing dataset and contains 55 outdoor scenes. The non-homogeneous haze has ... | Provide a detailed description of the following dataset: NH-HAZE |
PVDN | PVDN is a dataset of vehicle detection at night, using light reflections caused by their headlamps. It contains 59,746 annotated grayscale images out of 346 different scenes in a rural environment at night. In these images, all oncoming vehicles, their corresponding light objects (e. g., headlamps), and their respectiv... | Provide a detailed description of the following dataset: PVDN |
JEC-QA | JEC-QA is a LQA (Legal Question Answering) dataset collected from the National Judicial Examination of China. It contains 26,365 multiple-choice and multiple-answer questions in total. The task of the dataset is to predict the answer using the questions and relevant articles. To do well on JEC-QA, both retrieving and a... | Provide a detailed description of the following dataset: JEC-QA |
TrashCan | The TrashCan dataset is an instance-segmentation dataset of underwater trash. It is comprised of annotated images (7,212 images) which contain observations of trash, ROVs, and a wide variety of undersea flora and fauna. The annotations in this dataset take the format of instance segmentation annotations: bitmaps contai... | Provide a detailed description of the following dataset: TrashCan |
UCLA Aerial Event Dataset | The UCLA Aerial Event Dataest has been captured by a low-cost hex-rotor with a GoPro camera, which is able to eliminate the high frequency vibration of the camera and hold in air autonomously through a GPS and a barometer. It can also fly 20 ∼ 90m above the ground and stays 5 minutes in air.
This hex-rotor has been... | Provide a detailed description of the following dataset: UCLA Aerial Event Dataset |
CQASUMM | CQASUMM is a dataset for CQA (Community Question Answering) summarization, constructed from the 4.4 million Yahoo! Answers L6 dataset. The dataset contains ~300k annotated samples. | Provide a detailed description of the following dataset: CQASUMM |
NeuralNews | NeuralNews is a dataset for machine-generated news detection. It consists of human-generated and machine-generated articles. The human-generated articles are extracted from the GoodNews dataset, which is extracted from the New York Times. It contains 4 types of articles:
- Real Articles and Real Captions
- Real Art... | Provide a detailed description of the following dataset: NeuralNews |
EyeCar | EyeCar is a dataset of driving videos of vehicles involved in rear-end collisions paired with eye fixation data captured from human subjects. It contains 21 front-view videos that were captured in various traffic, weather, and day light conditions. Each video is 30sec in length and contains typical driving tasks (e.g.,... | Provide a detailed description of the following dataset: EyeCar |
WordNet-feelings | WordNet-feelings, is an affective dataset that identifies 3664 word senses as feelings, and associates each of these with one of the 9 categories of feeling. The 9 different categories are: Actions, Anger, Attention, Attraction, Hedonics, Other, Physiological, Social, Wellbeing. | Provide a detailed description of the following dataset: WordNet-feelings |
Doc3DShade | Doc3DShade extends Doc3D with realistic lighting and shading. Follows a similar synthetic rendering procedure using captured document 3D shapes but final image generation step combines real shading of different types of paper materials under numerous illumination conditions. | Provide a detailed description of the following dataset: Doc3DShade |
Deep Fakes Dataset | The Deep Fakes Dataset is a collection of "in the wild" portrait videos for deepfake detection. The videos in the dataset are diverse real-world samples in terms of the source generative model, resolution, compression, illumination, aspect-ratio, frame rate, motion, pose, cosmetics, occlusion, content, and context. The... | Provide a detailed description of the following dataset: Deep Fakes Dataset |
GSL | ## Dataset Description
The [Greek Sign Language (GSL)](https://arxiv.org/abs/2007.12530) is a large-scale RGB+D dataset, suitable for Sign Language Recognition (SLR) and Sign Language Translation (SLT). The video captures are conducted using an Intel RealSense D435 RGB+D camera at a rate of 30 fps. Both the RGB and... | Provide a detailed description of the following dataset: GSL |
SketchyScene | SketchyScene is a large-scale dataset of scene sketches to advance research on sketch understanding at both the object and scene level. The dataset is created through a novel and carefully designed crowdsourcing pipeline, enabling users to efficiently generate large quantities of realistic and diverse scene sketches. S... | Provide a detailed description of the following dataset: SketchyScene |
ECHR | ECHR is an English legal judgment prediction dataset of cases from the European Court of Human Rights (ECHR). The dataset contains ~11.5k cases, including the raw text.
For each case, the dataset provides a list of facts extracted using regular expressions from the case description. Each case is also mapped to artic... | Provide a detailed description of the following dataset: ECHR |
AmazonQA | AmazonQA consists of 923k questions, 3.6M answers and 14M reviews across 156k products. Building on the well-known Amazon dataset, additional annotations are collected, marking each question as either answerable or unanswerable based on the available reviews. | Provide a detailed description of the following dataset: AmazonQA |
emrQA | emrQA has 1 million question-logical form and 400,000+ questionanswer evidence pairs. | Provide a detailed description of the following dataset: emrQA |
SuperGLUE | **SuperGLUE** is a benchmark dataset designed to pose a more rigorous test of language understanding than GLUE. SuperGLUE has the same high-level motivation as GLUE: to provide a simple, hard-to-game measure of progress toward general-purpose language understanding technologies for English. SuperGLUE follows the basic ... | Provide a detailed description of the following dataset: SuperGLUE |
TurkQA | TurkQA consists of a selection of sentences from English Wikipedia articles, with questions and answers crowdsourced from workers on Amazon Mechanical Turk. | Provide a detailed description of the following dataset: TurkQA |
XTREME | The **Cross-lingual TRansfer Evaluation of Multilingual Encoders (XTREME)** benchmark was introduced to encourage more research on multilingual transfer learning,. XTREME covers 40 typologically diverse languages spanning 12 language families and includes 9 tasks that require reasoning about different levels of syntax ... | Provide a detailed description of the following dataset: XTREME |
WikiMovies | WikiMovies is a dataset for question answering for movies content. It contains ~100k questions in the movie domain, and was designed to be answerable by using either a perfect KB (based on OMDb), | Provide a detailed description of the following dataset: WikiMovies |
MDD | Movie Dialog dataset (MDD) is designed to measure how well models can perform at goal and non-goal orientated dialog centered around the topic of movies (question answering, recommendation and discussion). | Provide a detailed description of the following dataset: MDD |
CBT | Children’s Book Test (CBT) is designed to measure directly how well language models can exploit wider linguistic context. The CBT is built from books that are freely available thanks to Project Gutenberg. | Provide a detailed description of the following dataset: CBT |
Dialog-based Language Learning dataset | Dialog-based Language Learning dataset is designed to measure how well models can perform at learning as a student given a teacher’s textual responses to the student’s answer (as well as potentially receiving an external real-valued reward signal). | Provide a detailed description of the following dataset: Dialog-based Language Learning dataset |
TyDiQA-GoldP | **TyDiQA** is the gold passage version of the Typologically Diverse Question Answering (TyDiWA) dataset, a benchmark for information-seeking question answering, which covers nine languages. The gold passage version is a simplified version of the primary task, which uses only the gold passage as context and excludes una... | Provide a detailed description of the following dataset: TyDiQA-GoldP |
WikiReading | WikiReading is a large-scale natural language understanding task and publicly-available dataset with 18 million instances. The task is to predict textual values from the structured knowledge base Wikidata by reading the text of the corresponding Wikipedia articles. The task contains a rich variety of challenging classi... | Provide a detailed description of the following dataset: WikiReading |
Tatoeba | The **Tatoeba** dataset consists of up to 1,000 English-aligned sentence pairs covering 122 languages.
Image Source: [https://arxiv.org/pdf/1812.10464v2.pdf](https://arxiv.org/pdf/1812.10464v2.pdf) | Provide a detailed description of the following dataset: Tatoeba |
DeeperForensics-1.0 | **DeeperForensics-1.0** represents the largest face forgery detection dataset by far, with 60,000 videos constituted by a total of 17.6 million frames, 10 times larger than existing datasets of the same kind. The full dataset includes 48,475 source videos and 11,000 manipulated videos. The source videos are collected o... | Provide a detailed description of the following dataset: DeeperForensics-1.0 |
WikiSuggest | To collect WikiSuggest, Google Suggest API is used to harvest natural language questions and submit them to Google Search. Whenever Google Search returns a box with a short answer from Wikipedia, an example from the question, answer, and the Wikipedia document are created. If the answer string is missing from the docum... | Provide a detailed description of the following dataset: WikiSuggest |
FineGym | **FineGym** is an action recognition dataset build on top of gymnasium videos. Compared to existing action recognition datasets, FineGym is distinguished in richness, quality, and diversity. In particular, it provides temporal annotations at both action and sub-action levels with a three-level semantic hierarchy. For e... | Provide a detailed description of the following dataset: FineGym |
Shmoop Corpus | Shmoop Corpus is a dataset of 231 stories that are paired with detailed multi-paragraph summaries for each individual chapter (7,234 chapters), where the summary is chronologically aligned with respect to the story chapter. From the corpus, a set of common NLP tasks are constructed, including Cloze-form question answer... | Provide a detailed description of the following dataset: Shmoop Corpus |
BookTest | BookTest is a new dataset similar to the popular Children’s Book Test (CBT), however more than 60 times larger. | Provide a detailed description of the following dataset: BookTest |
MovieNet | **MovieNet** is a holistic dataset for movie understanding. MovieNet contains 1,100 movies with a large amount of multi-modal data, e.g. trailers, photos, plot descriptions, etc.. Besides, different aspects of manual annotations are provided in MovieNet, including 1.1M characters with bounding boxes and identities, 42K... | Provide a detailed description of the following dataset: MovieNet |
DREAM | DREAM is a multiple-choice Dialogue-based REAding comprehension exaMination dataset. In contrast to existing reading comprehension datasets, DREAM is the first to focus on in-depth multi-turn multi-party dialogue understanding.
DREAM contains 10,197 multiple choice questions for 6,444 dialogues, collected from Engli... | Provide a detailed description of the following dataset: DREAM |
MessyTable | **MessyTable** features a large number of scenes with messy tables captured from multiple camera views. Each scene in this dataset is highly complex, containing multiple object instances that could be identical, stacked and occluded by other instances. The key challenge is to associate all instances given the RGB image... | Provide a detailed description of the following dataset: MessyTable |
MCTest | MCTest is a freely available set of stories and associated questions intended for research on the machine comprehension of text.
MCTest requires machines to answer multiple-choice reading comprehension questions about fictional stories, directly tackling the high-level goal of open-domain machine comprehension. | Provide a detailed description of the following dataset: MCTest |
TweetQA | With social media becoming increasingly popular on which lots of news and real-time events are reported, developing automated question answering systems is critical to the effectiveness of many applications that rely on real-time knowledge. While previous question answering (QA) datasets have concentrated on formal tex... | Provide a detailed description of the following dataset: TweetQA |
UCF Sports | The UCF Sports dataset consists of a set of actions collected from various sports which are typically featured on broadcast television channels such as the BBC and ESPN. The video sequences were obtained from a wide range of stock footage websites including BBC Motion gallery and GettyImages.
The dataset includes a ... | Provide a detailed description of the following dataset: UCF Sports |
MSRA-B | The MSRA-B dataset is a dataset for salient object detection. It contains 5,000 images with a variety of image contents. Most of the images have a single salient object. There is a large variation among images including natural scenes, animals, indoor, outdoor, etc. | Provide a detailed description of the following dataset: MSRA-B |
VOT2015 | VOT2015 is a visual object tracking dataset. The dataset comprises 60 short sequences showing various objects in challenging backgrounds. The sequences were chosen from a large pool of sequences from different sources. | Provide a detailed description of the following dataset: VOT2015 |
VOT2014 | The dataset comprises 25 short sequences showing various objects in challenging backgrounds. Eight sequences are from the VOT2013 challenge (bolt, bicycle, david, diving, gymnastics, hand, sunshade, woman). The new sequences show complementary objects and backgrounds, for example a fish underwater or a surfer riding a ... | Provide a detailed description of the following dataset: VOT2014 |
UCF50 | UCF50 is an action recognition data set with 50 action categories, consisting of realistic videos taken from youtube. This data set is an extension of YouTube Action data set (UCF11) which has 11 action categories.
UCF50 data set's 50 action categories collected from youtube are: Baseball Pitch, Basketball Shooting,... | Provide a detailed description of the following dataset: UCF50 |
MSD | The Million Song Dataset is a freely-available collection of audio features and metadata for a million contemporary popular music tracks.
The core of the dataset is the feature analysis and metadata for one million songs, provided by The Echo Nest. The dataset does not include any audio, only the derived features. N... | Provide a detailed description of the following dataset: MSD |
CASME II | The Chinese Academy of Sciences Micro-Expression dataset (CASME II) consists of 255 videos, elicited from 26 participants. The videos are recorded using Point Gray GRAS-03K2C camera which has a frame rate of 200fps. The average video length is 0.34s, equivalent to 68 frames. Each video’s emotion label is annotated by t... | Provide a detailed description of the following dataset: CASME II |
UCF-Crime | The UCF-Crime dataset is a large-scale dataset of 128 hours of videos. It consists of 1900 long and untrimmed real-world surveillance videos, with 13 realistic anomalies including Abuse, Arrest, Arson, Assault, Road Accident, Burglary, Explosion, Fighting, Robbery, Shooting, Stealing, Shoplifting, and Vandalism. These ... | Provide a detailed description of the following dataset: UCF-Crime |
VOT2013 | The dataset comprises 16 short sequences showing various objects in challenging backgrounds. The sequences were chosen from a large pool of sequences using a methodology based on clustering visual features of object and background so that those 16 sequences sample evenly well the existing pool. The sequences were annot... | Provide a detailed description of the following dataset: VOT2013 |
Medical Segmentation Decathlon | The Medical Segmentation Decathlon is a collection of medical image segmentation datasets. It contains a total of 2,633 three-dimensional images collected across multiple anatomies of interest, multiple modalities and multiple sources. Specifically, it contains data for the following body organs or parts: Brain, Heart,... | Provide a detailed description of the following dataset: Medical Segmentation Decathlon |
CrossNER | CrossNER is a cross-domain NER (Named Entity Recognition) dataset, a fully-labeled collection of NER data spanning over five diverse domains (Politics, Natural Science, Music, Literature, and Artificial Intelligence) with specialized entity categories for different domains. Additionally, CrossNER also includes unlabele... | Provide a detailed description of the following dataset: CrossNER |
11k Hands | A large dataset of human hand images (dorsal and palmar sides) with detailed ground-truth information for gender recognition and biometric identification. | Provide a detailed description of the following dataset: 11k Hands |
2-PM Vessel Dataset | 2-PM Vessel is an open-source volumetric brain vasculature dataset obtained with two-photon microscopy at Focused Ultrasound Lab, at Sunnybrook Research Institute (affiliated with University of Toronto by Dr. Alison Burgess, Charissa Poon and Marc Santos. The dataset contains a total of 12 volumetric stacks consisting ... | Provide a detailed description of the following dataset: 2-PM Vessel Dataset |
Placepedia | **Placepedia** contains 240K places with 35M images from all over the world. Each place is associated with its district, city/town/village, state/province, country, continent, and a large amount of diverse photos. Both administrative areas and places have rich side information, e.g. discription, population, category, f... | Provide a detailed description of the following dataset: Placepedia |
2WikiMultiHopQA | Uses structured and unstructured data. The dataset introduces the evidence information containing a reasoning path for multi-hop questions. | Provide a detailed description of the following dataset: 2WikiMultiHopQA |
30MQA | An enormous question answer pair corpus produced by applying a novel neural network architecture on the knowledge base Freebase to transduce facts into natural language questions. | Provide a detailed description of the following dataset: 30MQA |
360-SOD | 360-SOD contains 500 high-resolution equirectangular images. | Provide a detailed description of the following dataset: 360-SOD |
3D60 | Collects high quality 360 datasets with ground truth depth annotations, by re-using recently released large scale 3D datasets and re-purposing them to 360 via rendering. | Provide a detailed description of the following dataset: 3D60 |
3D Hand Pose | **3D Hand Pose** is a multi-view hand pose dataset consisting of color images of hands and different kind of annotations for each: the bounding box and the 2D and 3D location on the joints in the hand. | Provide a detailed description of the following dataset: 3D Hand Pose |
3D Ken Burns Dataset | Provides a large-scale synthetic dataset which contains accurate ground truth depth of various photo-realistic scenes. | Provide a detailed description of the following dataset: 3D Ken Burns Dataset |
3DMAD | The 3D Mask Attack Database (3DMAD) is a biometric (face) spoofing database. It currently contains 76500 frames of 17 persons, recorded using Kinect for both real-access and spoofing attacks. Each frame consists of:
- a depth image (640x480 pixels – 1x11 bits)
- the corresponding RGB image (640x480 pixels – 3x8 bit... | Provide a detailed description of the following dataset: 3DMAD |
3DPeople Dataset | A large-scale synthetic dataset with 2.5 Million photo-realistic images of 80 subjects performing 70 activities and wearing diverse outfits. | Provide a detailed description of the following dataset: 3DPeople Dataset |
3DSeg-8 | The 3DSeg-8 is a collection of several publicly available 3D segmentation datasets from different medical imaging modalities, e.g. magnetic resonance imaging (MRI) and computed tomography (CT), with various scan regions, target organs and pathologies. | Provide a detailed description of the following dataset: 3DSeg-8 |
3D-ZeF | **3D-ZeF** dataset consists of eight sequences with a duration between 15-120 seconds and 1-10 free moving zebrafish. The videos have been annotated with a total of 86,400 points and bounding boxes. | Provide a detailed description of the following dataset: 3D-ZeF |
3RScan | A novel dataset and benchmark, which features 1482 RGB-D scans of 478 environments across multiple time steps. Each scene includes several objects whose positions change over time, together with ground truth annotations of object instances and their respective 6DoF mappings among re-scans. | Provide a detailed description of the following dataset: 3RScan |
4Seasons | 4Seasons is adataset covering seasonal and challenging perceptual conditions for autonomous driving. | Provide a detailed description of the following dataset: 4Seasons |
A2D2 | Audi Autonomous Driving Dataset (A2D2) consists of simultaneously recorded images and 3D point clouds, together with 3D bounding boxes, semantic segmentation, instance segmentation, and data extracted from the automotive bus. | Provide a detailed description of the following dataset: A2D2 |
A*3D | The **A*3D** dataset is a step forward to make autonomous driving safer for pedestrians and the public in the real world.
Characteristics:
* 230K human-labeled 3D object annotations in 39,179 LiDAR point cloud frames and corresponding frontal-facing RGB images.
* Captured at different times (day, night) and weathers (s... | Provide a detailed description of the following dataset: A*3D |
Aachen Day-Night | **Aachen Day-Night** is a dataset designed for benchmarking 6DOF outdoor visual localization in changing conditions. It focuses on localizing high-quality night-time images against a day-time 3D model. There are 14,607 images with changing conditions of weather, season and day-night cycles. | Provide a detailed description of the following dataset: Aachen Day-Night |
AADB | Contains aesthetic scores and meaningful attributes assigned to each image by multiple human raters. | Provide a detailed description of the following dataset: AADB |
AAVE/SAE Paired Dataset | AAVE/SAE Paired Dataset contains 2019 intent-equivalent AAVE/SAE pairs. The AAVE (African-American Vernacular English) samples are sampled from Blodgett et. al. (2016)'s TwitterAAE, with their corresponding SAE (Standard American English) samples annotated by Amazon MTurk. | Provide a detailed description of the following dataset: AAVE/SAE Paired Dataset |
ABC Dataset | The **ABC Dataset** is a collection of one million Computer-Aided Design (CAD) models for research of geometric deep learning methods and applications. Each model is a collection of explicitly parametrized curves and surfaces, providing ground truth for differential quantities, patch segmentation, geometric feature det... | Provide a detailed description of the following dataset: ABC Dataset |
ACL ARC | ACL Anthology Reference Corpus (ACL ARC) is a collection of 10,920 academic papers from the ACL Anthology. ACL ARC is cleaned to remove:
- files that look like not full papers, paper fragments, foreign-language papers (e.g., French), or pure junk.
- headers (title and author information; NOT abstract).
- footers (... | Provide a detailed description of the following dataset: ACL ARC |
ACRONYM | A dataset for robot grasp planning based on physics simulation. The dataset contains 17.7M parallel-jaw grasps, spanning 8872 objects from 262 different categories, each labeled with the grasp result obtained from a physics simulator. | Provide a detailed description of the following dataset: ACRONYM |
Acronym Identification | Is an acronym disambiguation (AD) dataset for scientific domain with 62,441 samples which is significantly larger than the previous scientific AD dataset. | Provide a detailed description of the following dataset: Acronym Identification |
ActioNet | **ActioNet** is a video task-based dataset collected in a synthetic 3D environment. It contains 3,038 annotated videos and hierarchical task structures over 65 individual household tasks from 120 different scenes. Each task is annotated across three to five different scenes by 10 different annotators. The tasks can be ... | Provide a detailed description of the following dataset: ActioNet |
ActivityNet Entities | ActivityNet-Entities, augments the challenging ActivityNet Captions dataset with 158k bounding box annotations, each grounding a noun phrase. This allows training video description models with this data, and importantly, evaluate how grounded or "true" such model are to the video they describe.
Source: [https://github... | Provide a detailed description of the following dataset: ActivityNet Entities |
ActivityNet-QA | The ActivityNet-QA dataset contains 58,000 human-annotated QA pairs on 5,800 videos derived from the popular ActivityNet dataset. The dataset provides a benchmark for testing the performance of VideoQA models on long-term spatio-temporal reasoning. | Provide a detailed description of the following dataset: ActivityNet-QA |
ActivityNet Thumbnails | Consists of 10,000+ video-sentence pairs with each accompanied by an annotated sentence specified video thumbnail. | Provide a detailed description of the following dataset: ActivityNet Thumbnails |
ADHA | ADHA: “Adverbs Describing Human Actions” is the first benchmark for a new problem — recognizing human action adverbs (HAA). This is the first step for computer vision to change over from pattern recognition to real AI. Some key features of ADHA are: a semantically complete set of adverbs describing human actions, a set... | Provide a detailed description of the following dataset: ADHA |
ADL Piano MIDI | The **ADL Piano MIDI** is a dataset of 11,086 piano pieces from different genres. This dataset is based on the Lakh MIDI dataset, which is a collection on 45,129 unique MIDI files that have been matched to entries in the Million Song Dataset. Most pieces in the Lakh MIDI dataset have multiple instruments, so for each f... | Provide a detailed description of the following dataset: ADL Piano MIDI |
Advice Seeking Questions | The Advice-Seeking Questions (ASQ) dataset is a collection of personal narratives with advice-seeking questions. The dataset has been split into train, test, heldout sets, with 8865, 2500, 10000 test instances each. This dataset is used to train and evaluate methods that can infer what is the advice-seeking goal behind... | Provide a detailed description of the following dataset: Advice Seeking Questions |
ADVIO | Provides a wide range of raw sensor data that is accessible on almost any modern-day smartphone together with a high-quality ground-truth track. | Provide a detailed description of the following dataset: ADVIO |
AeroRIT | AeroRIT is a hyperspectral dataset to facilitate aerial hyperspectral scene understanding. | Provide a detailed description of the following dataset: AeroRIT |
AESLC | To study the task of email subject line generation: automatically generating an email subject line from the email body. | Provide a detailed description of the following dataset: AESLC |
Aesthetics Text Corpus | An exhaustive list of stop lemmas created from 12 corpora across multiple domains, consisting of over 13 million words, from which more than 200,000 lemmas were generated, and 11 publicly available stop word lists comprising over 1000 words, from which nearly 400 unique lemmas were generated. | Provide a detailed description of the following dataset: Aesthetics Text Corpus |
Affective Text | Affective Text (Test Corpus of SemEval 2007) by [Carlo Strapparava & Rada Mihalcea](https://www.aclweb.org/anthology/S07-1013/). | Provide a detailed description of the following dataset: Affective Text |
Aff-Wild | Aff-Wild is a dataset for emotion recognition from facial images in a variety of head poses, illumination conditions and occlusions. | Provide a detailed description of the following dataset: Aff-Wild |
Aff-Wild2 | Aff-Wild2 is an extension of the Aff-Wild dataset for affect recognition. It approximately doubles the number of included video frames and the number of subjects; thus, improving the variability of the included behaviors and of the involved persons. | Provide a detailed description of the following dataset: Aff-Wild2 |
AfroMNIST | A set of synthetic MNIST-style datasets for four orthographies used in Afro-Asiatic and Niger-Congo languages: Ge`ez (Ethiopic), Vai, Osmanya, and N'Ko. These datasets serve as "drop-in" replacements for MNIST. | Provide a detailed description of the following dataset: AfroMNIST |
Agriculture-Vision | A large-scale aerial farmland image dataset for semantic segmentation of agricultural patterns. Collects 94,986 high-quality aerial images from 3,432 farmlands across the US, where each image consists of RGB and Near-infrared (NIR) channels with resolution as high as 10 cm per pixel. | Provide a detailed description of the following dataset: Agriculture-Vision |
AGRR-2019 | Consists of 7.5k sentences with gapping (as well as 15k relevant negative sentences) and comprises data from various genres: news, fiction, social media and technical texts. The dataset was prepared for the Automatic Gapping Resolution Shared Task for Russian (AGRR-2019) - a competition aimed at stimulating the develop... | Provide a detailed description of the following dataset: AGRR-2019 |
AI2D-RST | AI2D-RST is a multimodal corpus of 1000 English-language diagrams that represent topics in primary school natural sciences, such as food webs, life cycles, moon phases and human physiology. The corpus is based on the Allen Institute for Artificial Intelligence Diagrams (AI2D) dataset, a collection of diagrams with crow... | Provide a detailed description of the following dataset: AI2D-RST |
AIDER | Dataset aimed to do automated aerial scene classification of disaster events from on-board a UAV. | Provide a detailed description of the following dataset: AIDER |
AIRS | The **AIRS** (Aerial Imagery for Roof Segmentation) dataset provides a wide coverage of aerial imagery with 7.5 cm resolution and contains over 220,000 buildings. The task posed for AIRS is defined as roof segmentation. | Provide a detailed description of the following dataset: AIRS |
AirSim | **AirSim** is a simulator for drones, cars and more, built on Unreal Engine. It is open-source, cross platform, and supports software-in-the-loop simulation with popular flight controllers such as PX4 & ArduPilot and hardware-in-loop with PX4 for physically and visually realistic simulations. It is developed as an Unre... | Provide a detailed description of the following dataset: AirSim |
AISHELL-1 | AISHELL-1 is a corpus for speech recognition research and building speech recognition systems for Mandarin. | Provide a detailed description of the following dataset: AISHELL-1 |
AISHELL-2 | AISHELL-2 contains 1000 hours of clean read-speech data from iOS is free for academic usage. | Provide a detailed description of the following dataset: AISHELL-2 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.