dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
Oxford105k
**Oxford105k** is the combination of the Oxford5k dataset and 99782 negative images crawled from Flickr using 145 most popular tags. This dataset is used to evaluate search performance for object retrieval (reported as mAP) on a large scale.
Provide a detailed description of the following dataset: Oxford105k
DispScenes
The **DispScenes** dataset was created to address the specific problem of disparate image matching. The image pairs in all the datasets exhibit high levels of variation in illumination and viewpoint and also contain instances of occlusion. The DispScenes dataset provides manual ground truth keypoint correspondences for...
Provide a detailed description of the following dataset: DispScenes
Retrieval-SfM
The Retrieval-SFM dataset is used for instance image retrieval. The dataset contains 28559 images from 713 locations in the world. Each image has a label indicating the location it belongs to. Most locations are famous man-made architectures such as palaces and towers, which are relatively static and positively contrib...
Provide a detailed description of the following dataset: Retrieval-SfM
VGG Cell
The **VGG Cell** dataset (made up entirely of synthetic images) is the main public benchmark used to compare cell counting techniques. Source: [People, Penguins and Petri Dishes: Adapting Object Counting Models To New Visual Domains And Object Types Without Forgetting](https://arxiv.org/abs/1711.05586) Image Source: [...
Provide a detailed description of the following dataset: VGG Cell
Tiny Images
The image dataset TinyImages contains 80 million images of size 32×32 collected from the Internet, crawling the words in WordNet. **The authors have decided to withdraw it because it contains offensive content, and have asked the community to stop using it.**
Provide a detailed description of the following dataset: Tiny Images
Permuted MNIST
**Permuted MNIST** is an MNIST variant that consists of 70,000 images of handwritten digits from 0 to 9, where 60,000 images are used for training, and 10,000 images for test. The difference of this dataset from the original MNIST is that each of the ten tasks is the multi-class classification of a different random per...
Provide a detailed description of the following dataset: Permuted MNIST
MNIST-8M
MNIST8M is derived from the MNIST dataset by applying random deformations and translations to the dataset.
Provide a detailed description of the following dataset: MNIST-8M
SUN3D
**SUN3D** contains a large-scale RGB-D video database, with 8 annotated sequences. Each frame has a semantic segmentation of the objects in the scene and information about the camera pose. It is composed by 415 sequences captured in 254 different spaces, in 41 different buildings. Moreover, some places have been captur...
Provide a detailed description of the following dataset: SUN3D
TUM RGB-D
**TUM RGB-D** is an RGB-D dataset. It contains the color and depth images of a Microsoft Kinect sensor along the ground-truth trajectory of the sensor. The data was recorded at full frame rate (30 Hz) and sensor resolution (640x480). The ground-truth trajectory was obtained from a high-accuracy motion-capture system wi...
Provide a detailed description of the following dataset: TUM RGB-D
SceneNet
**SceneNet** is a dataset of labelled synthetic indoor scenes. There are several labeled indoor scenes, including: - 11 Bedroom scenes with 428 objects - 15 Office scenes with 1,203 objects - 11 Kitchen scenes with 797 objects - 10 Living Room scenes with 715 objects - 10 Bathrooms with 556 objects
Provide a detailed description of the following dataset: SceneNet
SceneNet RGB-D
SceneNet-RGBD is a synthetic dataset containing large-scale photorealistic renderings of indoor scene trajectories with pixel-level annotations. Random sampling permits virtually unlimited scene configurations, and the dataset creators provide a set of 5M rendered RGB-D images from over 15K trajectories in synthetic la...
Provide a detailed description of the following dataset: SceneNet RGB-D
SUN Attribute
The **SUN Attribute** dataset consists of 14,340 images from 717 scene categories, and each category is annotated with a taxonomy of 102 discriminate attributes. The dataset can be used for high-level scene understanding and fine-grained scene recognition.
Provide a detailed description of the following dataset: SUN Attribute
iSUN
**iSUN** is a ground truth of gaze traces on images from the SUN dataset. The collection is partitioned into 6,000 images for training, 926 for validation and 2,000 for test.
Provide a detailed description of the following dataset: iSUN
BMS-26
The **Berkeley Motion Segmentation** Dataset (**BMS-26**) is a dataset for motion segmentation, which consists of 26 video sequences with pixel-accurate segmentation annotation of moving objects. A total of 189 frames is annotated. 12 of the sequences are taken from the Hopkins 155 dataset and new annotation is added....
Provide a detailed description of the following dataset: BMS-26
Freiburg Groceries
**Freiburg Groceries** is a groceries classification dataset consisting of 5000 images of size 256x256, divided into 25 categories. It has imbalanced class sizes ranging from 97 to 370 images per class. Images were taken in various aspect ratios and padded to squares. Source: [XNAS: Neural Architecture Search with Exp...
Provide a detailed description of the following dataset: Freiburg Groceries
Freiburg Spatial Relations
The **Freiburg Spatial Relations** dataset features 546 scenes each containing two out of 25 household objects. The depicted spatial relations can roughly be described as on top, on top on the corner, inside, inside and inclined, next to, and inclined. The dataset contains the 25 object models as textured .obj and .dae...
Provide a detailed description of the following dataset: Freiburg Spatial Relations
Freiburg Street Crossing
The **Freiburg Street Crossing** dataset consists of data collected from three different street crossings in Freiburg, Germany; ; two of which were traffic light regulated intersections and one a zebra crossing without traffic lights. The data can be used to train agents to cross roads autonomously. Source: [http://ai...
Provide a detailed description of the following dataset: Freiburg Street Crossing
Freiburg Campus 3D Scan
The **Freiburg Campus 3D Scan** dataset consists of 3D area maps from the Freiburg campus that were scanned with 3D lasers. Areas include corridors, the outdoor campus, and some of the colleges and buildings. Source: [http://aisdatasets.informatik.uni-freiburg.de/streetcrossing/](http://aisdatasets.informatik.uni-frei...
Provide a detailed description of the following dataset: Freiburg Campus 3D Scan
Plant Centroids
**Plant Centroids** is a dataset for stem emerging points (SEP) detection in RGB and NIR image data. The dataset is meant to aid the construction of agricultural robots, where detecting SEPs is an important perception task (to position weeding or fertilizing tools at the plant’s center and finding natural landmarks in ...
Provide a detailed description of the following dataset: Plant Centroids
Freiburg Across Seasons
**Freiburg Across Seasons** captures long-term perceptual changes across a span of 3 years. Image sequences were recorded with a forward facing bumblebee stereo camera mounted on a car. During summer, the camera was mounted outside the car where as during winters the camera was inside the car. The image sequences are r...
Provide a detailed description of the following dataset: Freiburg Across Seasons
Freiburg Terrains
**Freiburg Terrains** consist of three parts: 3.7 hours of audio recordings of the microphone pointed at the robot wheels. It also contains 24K RGB images from the camera mounted on top of the robot. The dataset creators also provide the SLAM poses for each data collection run. The dataset can be used for terrain class...
Provide a detailed description of the following dataset: Freiburg Terrains
Freiburg Block Tasks
**Freiburg Block Tasks** is a dataset for robot skill learning. It consists of two datasets. The first data set consisted of three simulated robot tasks: stacking (A), color pushing (B) and color stacking (C). The data set contains 300 multi-view demonstration videos per task. The tasks are simulated with PyBullet. Of...
Provide a detailed description of the following dataset: Freiburg Block Tasks
Cityscapes-Motion
The **Cityscapes-Motion** dataset is a supplement to the semantic annotations provided by the Cityscapes dataset, containing 2975 training images and 500 validation images. The dataset creators provide manually annotated motion labels for the category of cars. The images are of resolution 2048×1024 pixels. The task to ...
Provide a detailed description of the following dataset: Cityscapes-Motion
KITTI-Motion
The **KITTI-Motion** dataset contains pixel-wise semantic class labels and moving object annotations for 255 images taken from the KITTI Raw dataset. The images are of resolution 1280×384 pixels and contain scenes of freeways, residential areas and inner-cities. The task is not just to semantically segment objects but ...
Provide a detailed description of the following dataset: KITTI-Motion
MobilityAids
**MobilityAids** is a dataset for perception of people and their mobility aids. The annotated dataset contains five classes: pedestrian, person in wheelchair, pedestrian pushing a person in a wheelchair, person using crutches and person using a walking frame. In total the hospital dataset has over 17, 000 annotated RGB...
Provide a detailed description of the following dataset: MobilityAids
RobotPush
**RobotPush** is a dataset for object singulation – the task of separating cluttered objects through physical interaction. The dataset contains 3456 training images with labels and 1024 validation images with labels. It consists of simulated and real-world data collected from a PR2 robot that equipped with a Kinect 2 c...
Provide a detailed description of the following dataset: RobotPush
DeepLocCross
**DeepLocCross** is a localization dataset that contains RGB-D stereo images captured at 1280 x 720 pixels at a rate of 20 Hz. The ground-truth pose labels are generated using a LiDAR-based SLAM system. In addition to the 6-DoF localization poses of the robot, the dataset additionally contains tracked detections of the...
Provide a detailed description of the following dataset: DeepLocCross
DeepLoc
**DeepLoc** is a large-scale urban outdoor localization dataset. The dataset is currently comprised of one scene spanning an area of 110 x 130 m, that a robot traverses multiple times with different driving patterns. The dataset creators use a LiDAR-based SLAM system with sub-centimeter and sub-degree accuracy to compu...
Provide a detailed description of the following dataset: DeepLoc
Freiburg Lighting Adaptable Map Tracking
**Freiburg Lighting Adaptable Map Tracking** is a dataset for camera trajectory estimation. The dataset consists of two subdatasets, each consisting of a Lighting Adaptable Map and three camera trajectories recorded under varying lighting conditions. The map meshes are stored in PLY format with custom properties and el...
Provide a detailed description of the following dataset: Freiburg Lighting Adaptable Map Tracking
Freiburg Poking
The **Freiburg Poking** dataset is a dataset for learning intuitive physics from physical interaction. It consists of 40K of interaction data with a KUKA LBR iiwa manipulator and a fixed Azure Kinect RGB-D camera. The dataset creators built an arena of styrofoam with walls for preventing objects from falling down. At a...
Provide a detailed description of the following dataset: Freiburg Poking
7-Scenes
The **7-Scenes** dataset is a collection of tracked RGB-D camera frames. The dataset may be used for evaluation of methods for different applications such as dense tracking and mapping and relocalization techniques. All scenes were recorded from a handheld Kinect RGB-D camera at 640×480 resolution. The dataset creators...
Provide a detailed description of the following dataset: 7-Scenes
Cross-Dataset Testbed
The Cross-dataset Testbed is a Decaf7 based cross-dataset image classification dataset, which contains 40 categories of images from 3 domains: 3,847 images in Caltech256, 4,000 images in ImageNet, and 2,626 images for SUN. In total there are 10,473 images of 40 categories from these three domains. Source: [Probability...
Provide a detailed description of the following dataset: Cross-Dataset Testbed
Washington RGB-D
**Washington RGB-D** is a widely used testbed in the robotic community, consisting of 41,877 RGB-D images organized into 300 instances divided in 51 classes of common indoor objects (e.g. scissors, cereal box, keyboard etc). Each object instance was positioned on a turntable and captured from three different viewpoints...
Provide a detailed description of the following dataset: Washington RGB-D
TUM Kitchen
The **TUM Kitchen** dataset is an action recognition dataset that contains 20 video sequences captured by 4 cameras with overlapping views. The camera network captures the scene from four viewpoints with 25 fps, and every RGB frame is of the resolution 384×288 by pixels. The action labels are frame-wise, and provided f...
Provide a detailed description of the following dataset: TUM Kitchen
HIC
The Hands in action dataset (**HIC**) dataset has RGB-D sequences of hands interacting with objects. Source: [Learning joint reconstruction of hands and manipulated objects](https://arxiv.org/abs/1904.05767) Image Source: [http://files.is.tue.mpg.de/dtzionas/Hand-Object-Capture/](http://files.is.tue.mpg.de/dtzionas/Ha...
Provide a detailed description of the following dataset: HIC
George Washington
The **George Washington** dataset contains 20 pages of letters written by George Washington and his associates in 1755 and thereby categorized into historical collection. The images are annotated at word level and contain approximately 5,000 words.
Provide a detailed description of the following dataset: George Washington
Watch-n-Patch
The **Watch-n-Patch** dataset was created with the focus on modeling human activities, comprising multiple actions in a completely unsupervised setting. It is collected with Microsoft Kinect One sensor for a total length of about 230 minutes, divided in 458 videos. 7 subjects perform human daily activities in 8 offices...
Provide a detailed description of the following dataset: Watch-n-Patch
Parzival
The **Parzival** dataset consists of 47 pages by three writers. These pages were taken from a medieval German manuscript from the 13th century that contains the epic poem Parzival by Wolfram von Eschenbach. The image size is 2000 x 3000 pixels. 24 pages are selected as training set; 14 pages are selected as test set; 2...
Provide a detailed description of the following dataset: Parzival
CDTB
Source: [https://www.vicos.si/Projects/CDTB 4.2 State-of-the-art Comparison A TH CTB (color-and-depth visual object tracking) dataset is recorded by several passive and active RGB-D setups and contains indoor as well as outdoor sequences acquired in direct sunlight. The sequences were recorded to contain significant ...
Provide a detailed description of the following dataset: CDTB
EgoDexter
The **EgoDexter** dataset provides both 2D and 3D pose annotations for 4 testing video sequences with 3190 frames. The videos are recorded with body-mounted camera from egocentric viewpoints and contain cluttered backgrounds, fast camera motion, and complex interactions with various objects. Fingertip positions were ma...
Provide a detailed description of the following dataset: EgoDexter
SynthHands
The **SynthHands** dataset is a dataset for hand pose estimation which consists of real captured hand motion retargeted to a virtual hand with natural backgrounds and interactions with different objects. The dataset contains data for male and female hands, both with and without interaction with objects. While the hand ...
Provide a detailed description of the following dataset: SynthHands
Washington RGB-D Scenes v2
The RGB-D Scenes Dataset v2 consists of 14 scenes containing furniture (chair, coffee table, sofa, table) and a subset of the objects in the RGB-D Object Dataset (bowls, caps, cereal boxes, coffee mugs, and soda cans). Each scene is a point cloud created by aligning a set of video frames using Patch Volumes Mapping.
Provide a detailed description of the following dataset: Washington RGB-D Scenes v2
Washington RGB-D Scenes
The RGB-D Scenes Dataset contains 8 scenes annotated with objects that belong to the Washington RGB-D Object Dataset. Each scene is a single video sequence consisting of multiple RGB-D frames. Source: [https://rgbd-dataset.cs.washington.edu/dataset/rgbd-scenes-v2/](https://rgbd-dataset.cs.washington.edu/dataset/rgbd-s...
Provide a detailed description of the following dataset: Washington RGB-D Scenes
MannequinChallenge
The **MannequinChallenge** Dataset (MQC) provides in-the-wild videos of people in static poses while a hand-held camera pans around the scene. The dataset consists of three splits for training, validation and testing.
Provide a detailed description of the following dataset: MannequinChallenge
Freiburg RGB-D People
The **Freiburg RGB-D People** dataset contains 3000+ RGB-D frames acquired in a university hall from three vertically mounted Kinect sensors. The data contains mostly upright walking and standing persons seen from different orientations and with different levels of occlusions. Source: [http://www2.informatik.uni-freib...
Provide a detailed description of the following dataset: Freiburg RGB-D People
Fraunhofer IPA Bin-Picking
The **Fraunhofer IPA Bin-Picking** dataset is a large-scale dataset comprising both simulated and real-world scenes for various objects (potentially having symmetries) and is fully annotated with 6D poses. A pyhsics simulation is used to create scenes of many parts in bulk by dropping objects in a random position and o...
Provide a detailed description of the following dataset: Fraunhofer IPA Bin-Picking
PAVIS RGB-D
**PAVIS RGB-D** is a dataset for person re-identification using depth information. The main motivation is that techniques such as SDALF fail when the individuals change their clothing, therefore they cannot be used for long-term video surveillance. Depth information is the solution to deal with this problem because it...
Provide a detailed description of the following dataset: PAVIS RGB-D
Couples Therapy
The **Couples Therapy** corpus contains audio, video recordings and manual transcriptions of conversations between 134 real-life couples attending marital therapy. In each session, one person selected a topic that was discussed over 10 minutes with the spouse. At the end of the session, both speakers were rated separat...
Provide a detailed description of the following dataset: Couples Therapy
Raider
The **Raider** dataset collects fMRI recordings of 1000 voxels from the ventral temporal cortex, for 10 healthy adult participants passively watching the full-length movie “Raiders of the Lost Ark”. Source: [Time-Resolved fMRI Shared Response Model using Gaussian Process Factor Analysis](https://arxiv.org/abs/2006.055...
Provide a detailed description of the following dataset: Raider
VizDoom
ViZDoom is an AI research platform based on the classical First Person Shooter game Doom. The most popular game mode is probably the so-called Death Match, where several players join in a maze and fight against each other. After a fixed time, the match ends and all the players are ranked by the FRAG scores defined as k...
Provide a detailed description of the following dataset: VizDoom
StarCraft II Learning Environment
The **StarCraft II Learning Environment** (S2LE) is a reinforcement learning environment based on the game StarCraft II. The environment consists of three sub-components: a Linux StarCraft II binary, the StarCraft II API and PySC2. The StarCraft II API allows programmatic control of StarCraft II. It can be used to star...
Provide a detailed description of the following dataset: StarCraft II Learning Environment
AI2-THOR
AI2-Thor is an interactive environment for embodied AI. It contains four types of scenes, including kitchen, living room, bedroom and bathroom, and each scene includes 30 rooms, where each room is unique in terms of furniture placement and item types. There are over 2000 unique objects for AI agents to interact with.
Provide a detailed description of the following dataset: AI2-THOR
TORCS
**TORCS** (**The Open Racing Car Simulator**) is a driving simulator. It is capable of simulating the essential elements of vehicular dynamics such as mass, rotational inertia, collision, mechanics of suspensions, links and differentials, friction and aerodynamics. Physics simulation is simplified and is carried out th...
Provide a detailed description of the following dataset: TORCS
DeepMind Control Suite
The **DeepMind Control Suite** (DMCS) is a set of simulated continuous control environments with a standardized structure and interpretable rewards. The tasks are written and powered by the MuJoCo physics engine, making them easy to identify. Control Suite tasks include Pendulum, Acrobot, Cart-pole, Cart-k-pole, Ball i...
Provide a detailed description of the following dataset: DeepMind Control Suite
GVGAI
The **General Video Game AI** (**GVGAI**) framework is widely used in research which features a corpus of over 100 single-player games and 60 two-player games. These are fairly small games, each focusing on specific mechanics or skills the players should be able to demonstrate, including clones of classic arcade games ...
Provide a detailed description of the following dataset: GVGAI
StarData
**StarData** is a StarCraft: Brood War replay dataset, with 65,646 games. The full dataset after compression is 365 GB, 1535 million frames, and 496 million player actions. The entire frame data was dumped out at 8 frames per second.
Provide a detailed description of the following dataset: StarData
Atari-HEAD
**Atari-HEAD** is a dataset of human actions and eye movements recorded while playing Atari videos games. For every game frame, its corresponding image frame, the human keystroke action, the reaction time to make that action, the gaze positions, and immediate reward returned by the environment were recorded. The gaze d...
Provide a detailed description of the following dataset: Atari-HEAD
Mario AI
**Mario AI** was a benchmark environment for reinforcement learning. The gameplay in Mario AI, as in the original Nintendo’s version, consists in moving the controlled character, namely Mario, through two-dimensional levels, which are viewed sideways. Mario can walk and run to the right and left, jump, and (depending o...
Provide a detailed description of the following dataset: Mario AI
D4RL
**D4RL** is a collection of environments for offline reinforcement learning. These environments include Maze2D, AntMaze, Adroit, Gym, Flow, FrankKitchen and CARLA.
Provide a detailed description of the following dataset: D4RL
AtariARI
The **AtariARI** (**Atari Annotated RAM Interface**) is an environment for representation learning. The Atari Arcade Learning Environment (ALE) does not explicitly expose any ground truth state information. However, ALE does expose the RAM state (128 bytes per timestep) which are used by the game programmer to store im...
Provide a detailed description of the following dataset: AtariARI
Lani
LANI is a 3D navigation environment and corpus, where an agent navigates between landmarks. **Lani** contains 27,965 crowd-sourced instructions for navigation in an open environment. Each datapoint includes an instruction, a human-annotated ground-truth demonstration trajectory, and an environment with various landmark...
Provide a detailed description of the following dataset: Lani
CHALET
**CHALET** is a 3D house simulator with support for navigation and manipulation. Unlike existing systems, CHALET supports both a wide range of object manipulation, as well as supporting complex environemnt layouts consisting of multiple rooms. The range of object manipulations includes the ability to pick up and place ...
Provide a detailed description of the following dataset: CHALET
Griddly
**Griddly** is an environment for grid-world based research. Griddly provides a highly optimized game state and rendering engine with a flexible high-level interface for configuring environments. Not only does Griddly offer simple interfaces for single, multi-player and RTS games, but also multiple methods of renderin...
Provide a detailed description of the following dataset: Griddly
NomBank
**NomBank** is an annotation project at New York University that is related to the PropBank project at the University of Colorado. The goal is to mark the sets of arguments that cooccur with nouns in the PropBank Corpus (the Wall Street Journal Corpus of the Penn Treebank), just as PropBank records such information fo...
Provide a detailed description of the following dataset: NomBank
QA-SRL
**QA-SRL** was proposed as an open schema for semantic roles, in which the relation between an argument and a predicate is expressed as a natural-language question containing the predicate (“Where was someone educated?”) whose answer is the argument (“Princeton”). The authors collected about 19,000 question-answer pair...
Provide a detailed description of the following dataset: QA-SRL
SParC
**SParC** is a large-scale dataset for complex, cross-domain, and context-dependent (multi-turn) semantic parsing and text-to-SQL task (interactive natural language interfaces for relational databases).
Provide a detailed description of the following dataset: SParC
CoNLL 2002
The shared task of CoNLL-2002 concerns language-independent named entity recognition. The types of named entities include: persons, locations, organizations and names of miscellaneous entities that do not belong to the previous three groups. The participants of the shared task were offered training and test data for at...
Provide a detailed description of the following dataset: CoNLL 2002
Panlex
PanLex translates words in thousands of languages. Its database is panlingual (emphasizes coverage of every language) and lexical (focuses on words, not sentences).
Provide a detailed description of the following dataset: Panlex
MCScript
**MCScript** is used as the official dataset of SemEval2018 Task11. This dataset constructs a collection of text passages about daily life activities and a series of questions referring to each passage, and each question is equipped with two answer choices. The MCScript comprises 9731, 1411, and 2797 questions in train...
Provide a detailed description of the following dataset: MCScript
KP20k
**KP20k** is a large-scale scholarly articles dataset with 528K articles for training, 20K articles for validation and 20K articles for testing.
Provide a detailed description of the following dataset: KP20k
Semantic Scholar
The **Semantic Scholar** corpus (S2) is composed of titles from scientific papers published in machine learning conferences and journals from 1985 to 2017, split by year (33 timesteps).
Provide a detailed description of the following dataset: Semantic Scholar
EVALution
**EVALution** dataset is evenly distributed among the three classes (hypernyms, co-hyponyms and random) and involves three types of parts of speech (noun, verb, adjective). The full dataset contains a total of 4,263 distinct terms consisting of 2,380 nouns, 958 verbs and 972 adjectives.
Provide a detailed description of the following dataset: EVALution
Senseval-2
There are now many computer programs for automatically determining the sense of a word in context (Word Sense Disambiguation or WSD). The purpose of SENSEVAL is to evaluate the strengths and weaknesses of such programs with respect to different words, different varieties of language, and different languages.
Provide a detailed description of the following dataset: Senseval-2
RoboCup
**RoboCup** is an initiative in which research groups compete by enabling their robots to play football matches. Playing football requires solving several challenging tasks, such as vision, motion, and team coordination. Framing the research efforts onto football attracts public interest (and potential research funding...
Provide a detailed description of the following dataset: RoboCup
ShARC
**ShARC** is a Conversational Question Answering dataset focussing on question answering from texts containing rules.
Provide a detailed description of the following dataset: ShARC
SIQA
**Social Interaction QA (SIQA)** is a question-answering benchmark for testing social commonsense intelligence. Contrary to many prior benchmarks that focus on physical or taxonomic knowledge, Social IQa focuses on reasoning about people’s actions and their social implications. For example, given an action like "Jesse ...
Provide a detailed description of the following dataset: SIQA
OLID
The **OLID** is a hierarchical dataset to identify the type and the target of offensive texts in social media. The dataset is collected on Twitter and publicly available. There are 14,100 tweets in total, in which 13,240 are in the training set, and 860 are in the test set. For each tweet, there are three levels of lab...
Provide a detailed description of the following dataset: OLID
Multi-News
**Multi-News**, consists of news articles and human-written summaries of these articles from the site newser.com. Each summary is professionally written by editors and includes links to the original articles cited.
Provide a detailed description of the following dataset: Multi-News
CLOTH
The Cloze Test by Teachers (**CLOTH**) benchmark is a collection of nearly 100,000 4-way multiple-choice cloze-style questions from middle- and high school-level English language exams, where the answer fills a blank in a given text. Each question is labeled with a type of deep reasoning it involves, where the four pos...
Provide a detailed description of the following dataset: CLOTH
CosmosQA
CosmosQA is a large-scale dataset of 35.6K problems that require commonsense-based reading comprehension, formulated as multiple-choice questions. It focuses on reading between the lines over a diverse collection of people’s everyday narratives, asking questions concerning on the likely causes or effects of events that...
Provide a detailed description of the following dataset: CosmosQA
WinoBias
**WinoBias** contains 3,160 sentences, split equally for development and test, created by researchers familiar with the project. Sentences were created to follow two prototypical templates but annotators were encouraged to come up with scenarios where entities could be interacting in plausible ways. Templates were sele...
Provide a detailed description of the following dataset: WinoBias
Spades
Datasets **Spades** contains 93,319 questions derived from clueweb09 sentences. Specifically, the questions were created by randomly removing an entity, thus producing sentence-denotation pairs. Source: [Learning an Executable Neural Semantic Parser](https://arxiv.org/abs/1711.05066) Image Source: [https://github.com/...
Provide a detailed description of the following dataset: Spades
WikiSum
**WikiSum** is a dataset based on English Wikipedia and suitable for a task of multi-document abstractive summarization. In each instance, the input is comprised of a Wikipedia topic (title of article) and a collection of non-Wikipedia reference documents, and the target is the Wikipedia article text. The dataset is re...
Provide a detailed description of the following dataset: WikiSum
DRCD
Delta Reading Comprehension Dataset (DRCD) is an open domain traditional Chinese machine reading comprehension (MRC) dataset. This dataset aimed to be a standard Chinese machine reading comprehension dataset, which can be a source dataset in transfer learning. The dataset contains 10,014 paragraphs from 2,108 Wikipedia...
Provide a detailed description of the following dataset: DRCD
EmotionLines
**EmotionLines** contains a total of 29245 labeled utterances from 2000 dialogues. Each utterance in dialogues is labeled with one of seven emotions, six Ekman’s basic emotions plus the neutral emotion. Each labeling was accomplished by 5 workers, and for each utterance in a label, the emotion category with the highest...
Provide a detailed description of the following dataset: EmotionLines
Chinese Gigaword
**Chinese Gigaword** corpus consists of 2.2M of headline-document pairs of news stories covering over 284 months from two Chinese newspapers, namely the Xinhua News Agency of China (XIN) and the Central News Agency of Taiwan (CNA).
Provide a detailed description of the following dataset: Chinese Gigaword
CELEX
**CELEX** database comprises three different searchable lexical databases, Dutch, English and German. The lexical data contained in each database is divided into five categories: orthography, phonology, morphology, syntax (word class) and word frequency.
Provide a detailed description of the following dataset: CELEX
MuST-C
**MuST-C** currently represents the largest publicly available multilingual corpus (one-to-many) for speech translation. It covers eight language directions, from English to German, Spanish, French, Italian, Dutch, Portuguese, Romanian and Russian. The corpus consists of audio, transcriptions and translations of Englis...
Provide a detailed description of the following dataset: MuST-C
Who-did-What
**Who-did-What** collects its corpus from news and provides options for questions similar to CBT. Each question is formed from two independent articles: an article is treated as context to be read and a separate article on the same event is used to form the query.
Provide a detailed description of the following dataset: Who-did-What
MetaQA
The **MetaQA** dataset consists of a movie ontology derived from the WikiMovies Dataset and three sets of question-answer pairs written in natural language: 1-hop, 2-hop, and 3-hop queries.
Provide a detailed description of the following dataset: MetaQA
FakeNewsNet
**FakeNewsNet** is collected from two fact-checking websites: GossipCop and PolitiFact containing news contents with labels annotated by professional journalists and experts, along with social context information.
Provide a detailed description of the following dataset: FakeNewsNet
STS 2014
STS-2014 is from SemEval-2014, constructed from image descriptions, news headlines, tweet news, discussion forums, and OntoNotes.
Provide a detailed description of the following dataset: STS 2014
MEDIA
The **MEDIA** French corpus is dedicated to semantic extraction from speech in a context of human/machine dialogues. The corpus has manual transcription and conceptual annotation of dialogues from 250 speakers. It is split into the following three parts : (1) the training set (720 dialogues, 12K sentences), (2) the dev...
Provide a detailed description of the following dataset: MEDIA
ASPEC
**ASPEC**, Asian Scientific Paper Excerpt Corpus, is constructed by the Japan Science and Technology Agency (JST) in collaboration with the National Institute of Information and Communications Technology (NICT). It consists of a Japanese-English paper abstract corpus of 3M parallel sentences (ASPEC-JE) and a Japanese-C...
Provide a detailed description of the following dataset: ASPEC
OMICS
**OMICS** is an extensive collection of knowledge for indoor service robots gathered from internet users. Currently, it contains 48 tables capturing different sorts of knowledge. Each tuple of the Help table maps a user desire to a task that may meet the desire (e.g., ⟨ “feel thirsty”, “by offering drink” ⟩). Each tupl...
Provide a detailed description of the following dataset: OMICS
QUASAR
The Question Answering by Search And Reading (**QUASAR**) is a large-scale dataset consisting of [QUASAR-S](quasar-s) and [QUASAR-T](quasar-t). Each of these datasets is built to focus on evaluating systems devised to understand a natural language query, a large corpus of texts and to extract an answer to the question ...
Provide a detailed description of the following dataset: QUASAR
Dialogue State Tracking Challenge
The Dialog State Tracking Challenges 2 & 3 (DSTC2&3) were research challenge focused on improving the state of the art in tracking the state of spoken dialog systems. State tracking, sometimes called belief tracking, refers to accurately estimating the user's goal as a dialog progresses. Accurate state tracking is desi...
Provide a detailed description of the following dataset: Dialogue State Tracking Challenge
ISEAR
Over a period of many years during the 1990s, a large group of psychologists all over the world collected data in the **ISEAR** project, directed by Klaus R. Scherer and Harald Wallbott. Student respondents, both psychologists and non-psychologists, were asked to report situations in which they had experienced all of 7...
Provide a detailed description of the following dataset: ISEAR
CMRC
CMRC is a dataset is annotated by human experts with near 20,000 questions as well as a challenging set which is composed of the questions that need reasoning over multiple clues.
Provide a detailed description of the following dataset: CMRC
PubMed RCT
**PubMed 200k RCT** is new dataset based on PubMed for sequential sentence classification. The dataset consists of approximately 200,000 abstracts of randomized controlled trials, totaling 2.3 million sentences. Each sentence of each abstract is labeled with their role in the abstract using one of the following classes...
Provide a detailed description of the following dataset: PubMed RCT