dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
Action Recognition in the Dark
ARID is a dataset for action recognition in dark videos. It consists of over 3,780 video clips with 11 action categories.
Provide a detailed description of the following dataset: Action Recognition in the Dark
Kuzushiji-Kanji
Kuzushiji-Kanji is an imbalanced dataset of total 3832 Kanji characters (64x64 grayscale, 140,426 images), ranging from 1,766 examples to only a single example per class. Kuzushiji is a Japanese cursive writing style.
Provide a detailed description of the following dataset: Kuzushiji-Kanji
UCO-LAEO
A dataset for building models that detect people Looking At Each Other (LAEO) in video sequences.
Provide a detailed description of the following dataset: UCO-LAEO
Skeletics 152
A curated and 3-D pose-annotated subset of RGB videos sourced from Kinetics-700, a large-scale action dataset.
Provide a detailed description of the following dataset: Skeletics 152
TVQA+
TVQA+ contains 310.8K bounding boxes, linking depicted objects to visual concepts in questions and answers.
Provide a detailed description of the following dataset: TVQA+
Spot the Difference Corpus
Spot the Difference Corpus is a corpus of task-oriented spontaneous dialogues which contains 54 interactions between pairs of subjects interacting to find differences in two very similar scenes. The corpus includes rich transcriptions, annotations, audio and video.
Provide a detailed description of the following dataset: Spot the Difference Corpus
CCPE-M
A dataset consisting of 502 English dialogs with 12,000 annotated utterances between a user and an assistant discussing movie preferences in natural language. The corpus was constructed from dialogues between two paid crowd-workers using a Wizard-of-Oz methodology. One worker plays the role of an "assistant", while ...
Provide a detailed description of the following dataset: CCPE-M
COCO-CN
COCO-CN is a bilingual image description dataset enriching MS-COCO with manually written Chinese sentences and tags. The new dataset can be used for multiple tasks including image tagging, captioning and retrieval, all in a cross-lingual setting.
Provide a detailed description of the following dataset: COCO-CN
T2 Guiding
T2 Guiding is a dataset of 1000 images, each with six image labels. The images are from the Open Images Dataset (OID) and the dataset includes 2 sets of machine-generated labels for these images. * Object labels: Three random object labels generated by a FRCNN model trained on Visual Genome. * Image labels: Three r...
Provide a detailed description of the following dataset: T2 Guiding
FarsBase-KBP
FarsBase-KBP contains 22015 sentences, in which the entities and relation types are linked to the FarsBase ontology. This gold dataset can be reused for benchmarking KBP systems in the Persian language.
Provide a detailed description of the following dataset: FarsBase-KBP
Visual Wake Words
Visual Wake Words represents a common microcontroller vision use-case of identifying whether a person is present in the image or not, and provides a realistic benchmark for tiny vision models.
Provide a detailed description of the following dataset: Visual Wake Words
ReQA
Retrieval Question-Answering (ReQA) benchmark tests a model’s ability to retrieve relevant answers efficiently from a large set of documents.
Provide a detailed description of the following dataset: ReQA
ROSE
Retinal OCTA SEgmentation dataset (ROSE) consists of 229 OCTA images with vessel annotations at either centerline-level or pixel level.
Provide a detailed description of the following dataset: ROSE
UDIVA
UDIVA is a new non-acted dataset of face-to-face dyadic interactions, where interlocutors perform competitive and collaborative tasks with different behavior elicitation and cognitive workload. The dataset consists of 90.5 hours of dyadic interactions among 147 participants distributed in 188 sessions, recorded using m...
Provide a detailed description of the following dataset: UDIVA
MuST-Cinema
MuST-Cinema is a Multilingual Speech-to-Subtitles corpus ideal for building subtitle-oriented machine and speech translation systems. It comprises audio recordings from English TED Talks, which are automatically aligned at the sentence level with their manual transcriptions and translations. MuST-Cinema was built b...
Provide a detailed description of the following dataset: MuST-Cinema
LIV360SV
The dataset contains 26,645, 360 degree, street-level images collected via cycling with a GoPro Fusion camera, recorded Jan 14th -- 18th 2020. 10,106 advertisements were identified and classified as food (1335), alcohol (217), gambling (149) and other (8405) (e.g., cars and broadband).
Provide a detailed description of the following dataset: LIV360SV
i3-video
The i3-video dataset contains "is-it-instructional" annotations for 6.4k videos from [Youtube-8M](https://paperswithcode.com/dataset/youtube-8m). The videos are considered to be instructional if they focus on real-world human actions accompanied by procedural language that explains what’s happening on screen in reasona...
Provide a detailed description of the following dataset: i3-video
Open Images V4
Open Images V4 offers large scale across several dimensions: 30.1M image-level labels for 19.8k concepts, 15.4M bounding boxes for 600 object classes, and 375k visual relationship annotations involving 57 classes. For object detection in particular, 15x more bounding boxes than the next largest datasets (15.4M boxes on...
Provide a detailed description of the following dataset: Open Images V4
CLaRO
CLaRO is a new dataset of 234 Competency Questions that had been processed automatically into 106 patterns. The coverage of CLaRO, with its 93 main templates and 41 linguistic variants, is about 90% for unseen questions.
Provide a detailed description of the following dataset: CLaRO
ImagiFilter
ImagiFilter focusses on photographic and/or natural images, a very common use-case in computer vision research. Annotations for coarse prediction are provided, i.e. photographic vs. non-photographic, and smaller fine-grained prediction tasks where the non-photographic class is broken down into five classes: maps, drawi...
Provide a detailed description of the following dataset: ImagiFilter
CROSS
Cross-Reference Omnidirectional Stitching IQA is a novel omnidirectional image dataset containing stitched images as well as dual-fisheye images captured from standard quarters of 0◦, 90◦ , 180◦ and 270◦. In this manner, when evaluating the quality of an image stitched from a pair of fisheye images (e.g., 0◦ and 180◦),...
Provide a detailed description of the following dataset: CROSS
BIRD
Blocksworld Image Reasoning Dataset (BIRD) contains images of wooden blocks in different configurations, and the sequence of moves to rearrange one configuration to the other.
Provide a detailed description of the following dataset: BIRD
Almawave-SLU
Almawave-SLU is the first Italian dataset for Spoken Language Understanding (SLU). It is derived through a semi-automatic procedure and is used as a benchmark of various open source and commercial systems.
Provide a detailed description of the following dataset: Almawave-SLU
EPIC30M
EPIC30M contains a subset of 26.2 millions tweets related to three general diseases, namely Ebola, Cholera and Swine Flu, and another subset of 4.7 millions tweets of six global epidemic outbreaks, including 2009 H1N1 Swine Flu, 2010 Haiti Cholera, 2012 Middle-East Respiratory Syndrome (MERS), 2013 West African Ebola, ...
Provide a detailed description of the following dataset: EPIC30M
The Spoken Wikipedia Corpora
The SWC is a corpus of aligned Spoken Wikipedia articles from the English, German, and Dutch Wikipedia. This corpus has several outstanding characteristics: - hundreds of hours of aligned audio - from a diverse set of readers - about a diverse set of topics - in a well-researched textual genre - licensed under a...
Provide a detailed description of the following dataset: The Spoken Wikipedia Corpora
Flickr Audio Caption Corpus
The Flickr 8k Audio Caption Corpus contains 40,000 spoken captions of 8,000 natural images. It was collected in 2015 to investigate multimodal learning schemes for unsupervised speech pattern discovery. For a description of the corpus, see: D. Harwath and J. Glass, "Deep Multimodal Semantic Embeddings for Speech and...
Provide a detailed description of the following dataset: Flickr Audio Caption Corpus
PCVC
The **Persian Consonant Vowel Combination (PCVC)** dataset is a phoneme based speech dataset, and also the first free Persian speech dataset to help Persian speech researchers. This dataset contains of 23 Persian consonants and 6 vowels. The sound samples are all possible combinations of vowels and consonants (138 samp...
Provide a detailed description of the following dataset: PCVC
DensePose-Track
DensePose-Track is a dataset of videos where selected frames are annotated in the traditional DensePose manner.
Provide a detailed description of the following dataset: DensePose-Track
2000 HUB5 English
**2000 HUB5 English Evaluation Transcripts** was developed by the Linguistic Data Consortium (LDC) and consists of transcripts of 40 English telephone conversations used in the 2000 HUB5 evaluation sponsored by NIST (National Institute of Standards and Technology). The Hub5 evaluation series focused on conversatio...
Provide a detailed description of the following dataset: 2000 HUB5 English
Parkinson Speech Dataset
**Parkinson Speech Dataset** is an audio dataset consisting of recordings of 20 Parkinson's Disease (PD) patients and 20 healthy subjects. From all subjects, multiple types of sound recordings (26) are taken. The goal is to classify which patients have Parkinson's.
Provide a detailed description of the following dataset: Parkinson Speech Dataset
MDID
The Multimodal Document Intent Dataset (MDID) is a dataset for computing author intent from multimodal data from Instagram. It contains 1,299 Instagram posts covering a variety of topics, annotated with labels from three taxonomies. The samples are labelled with 7 labels of intent: Provocative, Informative, Advocative,...
Provide a detailed description of the following dataset: MDID
Ciona17
Ciona17 is a semantic segmentation dataset with pixel-level annotations pertaining to invasive species in a marine environment. Diverse outdoor illumination, a range of object shapes, colour, and severe occlusion provide a significant real world challenge for the computer vision community.
Provide a detailed description of the following dataset: Ciona17
Arabic Speech Corpus
The **Arabic Speech Corpus** (1.5 GB) is a Modern Standard Arabic (MSA) speech corpus for speech synthesis. The corpus contains phonetic and orthographic transcriptions of more than 3.7 hours of MSA speech aligned with recorded speech on the phoneme level. The annotations include word stress marks on the individual pho...
Provide a detailed description of the following dataset: Arabic Speech Corpus
Mivia Audio Events Dataset
The **MIVIA audio events** data set is composed of a total of 6000 events for surveillance applications, namely glass breaking, gun shots and screams. The 6000 events are divided into a training set (composed of 4200 events) and a test set (composed of 1800 events). In audio surveillance applications, the events of ...
Provide a detailed description of the following dataset: Mivia Audio Events Dataset
VRAI
VRAI is a large-scale vehicle ReID dataset for UAV-based intelligent applications. The dataset consists of 137, 613 images of 13, 022 vehicle instances. The images of each vehicle instance are captured by cameras of two DJI consumer UAVs at different locations, with a variety of view angles and flight-altitudes (15m to...
Provide a detailed description of the following dataset: VRAI
RWCP Sound Scene Database
The **RWCP Sound Scene Database** includes non-speech sounds recorded in an anechoic room, reconstructed signals in various rooms, impulse responses for a microphone array, speech data recorded with the same array, and recordings of background noises. It is intended for use when simulating sound scenes. It was develope...
Provide a detailed description of the following dataset: RWCP Sound Scene Database
UIT-ViIC
UIT-ViIC contains manually written captions for images from Microsoft COCO dataset relating to sports played with ball. UIT-ViIC consists of 19,250 Vietnamese captions for 3,850 images.
Provide a detailed description of the following dataset: UIT-ViIC
NAR
**NAR** is a dataset of audio recordings made with the humanoid robot Nao in real world conditions for sound recognition benchmarking. All the recordings were collected using the robot’s microphone and thus have the following characteristics: - recorded with low-quality sensors (300 Hz – 18 kHz bandpass) - suffering...
Provide a detailed description of the following dataset: NAR
IISc VINE
Indian Institute of Science VIdeo Naturalness Evaluation (IISc VINE) is a database consisting of 300 videos, obtained by applying different prediction models on different datasets, and accompanying human opinion scores.
Provide a detailed description of the following dataset: IISc VINE
MineNav
MinNav is a synthetic dataset based on the sandbox game Minecraft. The dataset uses several plug-in program to generate rendered image sequences with time-aligned depth maps, surface normal maps and camera poses. Thanks for the large game's community, there is an extremely large number of 3D open-world environment, use...
Provide a detailed description of the following dataset: MineNav
Minecraft House
**Minecraft House** is a crowd sourced dataset that collects examples of humans building houses in Minecraft. Each user is asked to build a CraftAssist: A Framework for Dialogue-enabled Interactive Agents house on a fixed time budget (30 minutes), without any additional guidance or instructions. Every action of the us...
Provide a detailed description of the following dataset: Minecraft House
ARVSU
ARVSU contains a vast body of image variations in visual scenes with an annotated utterance and a corresponding addressee for each scenario.
Provide a detailed description of the following dataset: ARVSU
m2cai16-tool-locations
The m2cai16-tool-locations dataset contains spatial tool annotations for 2,532 frames across the first 10 videos in the m2cai16-tool dataset, which includes 15 videos in total. The dataset consists of 3,141 annotations of 7 surgical instrument classes, with an average of 1.2 labels per frame and 7 instrument classes pe...
Provide a detailed description of the following dataset: m2cai16-tool-locations
Minecraft Segmentation
**Minecraft Segmentation** is a segmentation dataset for the [Minecraft House](https://www.paperswithcode.com/dataset/minecraft-house) that adds semantic segmentation labels for sub-components of the house. There are 2050 houses in total and 1038 distinct labels of subcomponents.
Provide a detailed description of the following dataset: Minecraft Segmentation
CryoNuSeg
CryoNuSeg is a fully annotated FS-derived cryosectioned and H&E-stained nuclei instance segmentation dataset. The dataset contains images from 10 human organs that were not exploited in other publicly available datasets, and is provided with three manual mark-ups to allow measuring intra-observer and inter-observer var...
Provide a detailed description of the following dataset: CryoNuSeg
Princeton Shape
The **Princeton Shape** dataset provides a repository of 3D models and software tools for evaluating shape-based retrieval and analysis algorithms. The motivation is to promote the use of standardized data sets and evaluation methods for research in matching, classification, clustering, and recognition of 3D models. ...
Provide a detailed description of the following dataset: Princeton Shape
Opusparcus
Opusparcus is a paraphrase corpus for six European languages: German, English, Finnish, French, Russian, and Swedish. The paraphrases are extracted from the OpenSubtitles2016 corpus, which contains subtitles from movies and TV shows. For each target language, the Opusparcus data have been partitioned into three type...
Provide a detailed description of the following dataset: Opusparcus
IKEA 3D
**IKEA 3D** is a dataset of IKEA 3D models and aligned images, which is suitable for pose estimation. There are 759 images and 219 models including Sketchup (skp) and Wavefront (obj) files.
Provide a detailed description of the following dataset: IKEA 3D
RSOC
RSOC is a large-scale object counting dataset with remote sensing images, which contains four important geographic objects: buildings, crowded ships in harbors, large-vehicles and small-vehicles in parking lots.
Provide a detailed description of the following dataset: RSOC
A Large Dataset of Object Scans
**A Large Dataset of Object Scans** is a dataset of more than ten thousand 3D scans of real objects. To create the dataset, the authors recruited 70 operators, equipped them with consumer-grade mobile 3D scanning setups, and paid them to scan objects in their environments. The operators scanned objects of their choosin...
Provide a detailed description of the following dataset: A Large Dataset of Object Scans
RGRS
RGRS is a dataset for collaboratior recommendation on the ResearchGate academic social network. The data has been collected from Jan. 2019 to April 2019 and includes raw data of 3980 RG users.
Provide a detailed description of the following dataset: RGRS
ObjectNet3D
**ObjectNet3D** is a large scale database for 3D object recognition, named, that consists of 100 categories, 90,127 images, 201,888 objects in these images and 44,147 3D shapes. Objects in the images in the database are aligned with the 3D shapes, and the alignment provides both accurate 3D pose annotation and the clos...
Provide a detailed description of the following dataset: ObjectNet3D
Event-Stream Dataset
Event-Stream Dataset is a robotic grasping dataset with 91 objects.
Provide a detailed description of the following dataset: Event-Stream Dataset
PersonalDialog
PersonalDialog is a large-scale multi-turn dialogue dataset containing various traits from a large number of speakers. The dataset consists of 20.83M sessions and 56.25M utterances from 8.47M speakers. Each utterance is associated with a speaker who is marked with traits like Age, Gender, Location, Interest Tags, etc. ...
Provide a detailed description of the following dataset: PersonalDialog
Thingi10K
**Thingi10K** is a dataset of 3D-Printing Models. Specifically there are 10,000 models from featured “things” on thingiverse.com, suitable for testing 3D printing techniques such as structural analysis , shape optimization, or solid geometry operations.
Provide a detailed description of the following dataset: Thingi10K
CocoDoom
CocoDoom is a collection of pre-recorded data extracted from Doom gaming sessions along with annotations in the MS Coco format.
Provide a detailed description of the following dataset: CocoDoom
VOCASET
**VOCASET** is a 4D face dataset with about 29 minutes of 4D scans captured at 60 fps and synchronized audio. The dataset has 12 subjects and 480 sequences of about 3-4 seconds each with sentences chosen from an array of standard protocols that maximize phonetic diversity.
Provide a detailed description of the following dataset: VOCASET
MSAW
Multi-Sensor All Weather Mapping (MSAW) is a dataset and challenge, which features two collection modalities (both SAR and optical). The dataset and challenge focus on mapping and building footprint extraction using a combination of these data sources. MSAW covers 120 km^2 over multiple overlapping collects and is anno...
Provide a detailed description of the following dataset: MSAW
ADE-Affordance
ADE-Affordance is a new dataset that builds upon ADE20k, which contains annotations enabling such rich visual reasoning.
Provide a detailed description of the following dataset: ADE-Affordance
PISC
The People in Social Context (PISC) dataset is a dataset that focuses on social relationships. It consists of 22,670 images of 9 types of social relationships. It has annotations for the bounding boxes of all people, as well as the social relationship between all pairs of people in the images. In addition, it also cont...
Provide a detailed description of the following dataset: PISC
MINOS
**MINOS** is a simulator designed to support the development of multisensory models for goal-directed navigation in complex indoor environments. MINOS leverages large datasets of complex 3D environments and supports flexible configuration of multimodal sensor suites.
Provide a detailed description of the following dataset: MINOS
WIKIOG
WIKIOG is a public collection which consists of over 1.75 million document-outline pairs for research on the OG task.
Provide a detailed description of the following dataset: WIKIOG
SemanticUSL
SemanticUSL is a dataset for domain adaptation for LiDAR point cloud semantic segmentation. The dataset has the same data format and ontology as SemanticKITTI.
Provide a detailed description of the following dataset: SemanticUSL
Jericho
Jericho is a learning environment for man-made Interactive Fiction (IF) games.
Provide a detailed description of the following dataset: Jericho
3D-FRONT
**3D-FRONT** (3D Furnished Rooms with layOuts and semaNTics) is large-scale, and comprehensive repository of synthetic indoor scenes highlighted by professionally designed layouts and a large number of rooms populated by high-quality textured 3D models with style compatibility. From layout semantics down to texture det...
Provide a detailed description of the following dataset: 3D-FRONT
JParaCrawl
JParaCrawl is a parallel corpus for English-Japanese, for which the amount of publicly available parallel corpora is still limited. The parallel corpus was constructed by broadly crawling the web and automatically aligning parallel sentences. The corpus amassed over 8.7 million sentence pairs.
Provide a detailed description of the following dataset: JParaCrawl
3ThreeDWorld
**TDW** is a 3D virtual world simulation platform, utilizing state-of-the-art video game engine technology. A TDW simulation consists of two components: a) the Build, a compiled executable running on the Unity3D Engine, which is responsible for image rendering, audio synthesis and physics simulations; and b) the Contro...
Provide a detailed description of the following dataset: 3ThreeDWorld
MUSIC
The Multi-Spectral Imaging via Computed Tomography (MUSIC) dataset is a two-part (2D- and 3D spectral) open access dataset for advanced image analysis of spectral radiographic (x-ray) scans, their tomographic reconstruction and the detection of specific materials within such scans. The scans operate at a photon energy ...
Provide a detailed description of the following dataset: MUSIC
MuMu
MuMu is a new dataset of more than 31k albums classified into 250 genre classes.
Provide a detailed description of the following dataset: MuMu
FSVQA
Full-Sentence Visual Question Answering (FSVQA) dataset, consisting of nearly 1 million pairs of questions and full-sentence answers for images, built by applying a number of rule-based natural language processing techniques to original VQA dataset and captions in the MS COCO dataset.
Provide a detailed description of the following dataset: FSVQA
Taskmaster-1
**Taskmaster-1** is a dialog dataset consisting of 13,215 task-based dialogs in English, including 5,507 spoken and 7,708 written dialogs created with two distinct procedures. Each conversation falls into one of six domains: ordering pizza, creating auto repair appointments, setting up ride service, ordering movie tick...
Provide a detailed description of the following dataset: Taskmaster-1
RealEstate10K
**RealEstate10K** is a large dataset of camera poses corresponding to 10 million frames derived from about 80,000 video clips, gathered from about 10,000 YouTube videos. For each clip, the poses form a trajectory where each pose specifies the camera position and orientation along the trajectory. These poses are derived...
Provide a detailed description of the following dataset: RealEstate10K
Wikipedia Generation
**Wikipedia Generation** is a dataset for article generation from Wikipedia from references at the end of Wikipedia page and the top 10 search results for the Wikipedia topic.
Provide a detailed description of the following dataset: Wikipedia Generation
WildestFaces
WildestFaces is tailored to study cross-domain recognition under a variety of adverse conditions.
Provide a detailed description of the following dataset: WildestFaces
FAD
FAD is a dataset that have roughly 200,000 attribute labels for the above traits, for over 10,000 facial images.
Provide a detailed description of the following dataset: FAD
VQA 360°
VQA 360° is a dataset for visual question answering on 360° images containing around 17,000 real-world image-question-answer triplets for a variety of question types.
Provide a detailed description of the following dataset: VQA 360°
StreetStyle
StreetStyle is a large-scale dataset of photos of people annotated with clothing attributes, and use this dataset to train attribute classifiers via deep learning.
Provide a detailed description of the following dataset: StreetStyle
PHSPD
PHSPD is a home-grown polarization image dataset of various human shapes and poses.
Provide a detailed description of the following dataset: PHSPD
HARRISON
HARRISON dataset is a benchmark on hashtag recommendation for real world images in social networks. The HARRISON dataset is a realistic dataset, composed of 57,383 photos from Instagram and an average of 4.5 associated hashtags for each photo.
Provide a detailed description of the following dataset: HARRISON
OC20
**Open Catalyst 2020** is a dataset for catalysis in chemical engineering. Focusing on molecules that are important in renewable energy applications, the OC20 data set comprises over 1.3 million relaxations of molecular adsorptions onto surfaces, the largest data set of electrocatalyst structures to date.
Provide a detailed description of the following dataset: OC20
FSOD
Few-Shot Object Detection Dataset (FSOD) is a high-diverse dataset specifically designed for few-shot object detection and intrinsically designed to evaluate thegenerality of a model on novel categories.
Provide a detailed description of the following dataset: FSOD
DUS
The **Daimler Urban Segmentation Dataset** is a dataset for semantic segmentation. It consists of video sequences recorded in urban traffic. The dataset consists of 5000 rectified stereo image pairs with a resolution of 1024x440. 500 frames (every 10th frame of the sequence) come with pixel-level semantic class annotat...
Provide a detailed description of the following dataset: DUS
HumanAct12
**HumanAct12** is a new 3D human motion dataset adopted from the polar image and 3D pose dataset PHSPD, with proper temporal cropping and action annotating. Statistically, there are 1191 3D motion clips(and 90,099 poses in total) which are categorized into 12 action classes, and 34 fine-grained sub-classes. The action ...
Provide a detailed description of the following dataset: HumanAct12
Large Age-Gap
**Large Age-Gap (LAG)** is a dataset for face verification, The dataset contains 3,828 images of 1,010 celebrities. For each identity at least one child/young image and one adult/old image are present.
Provide a detailed description of the following dataset: Large Age-Gap
Interestingness
The **Interestingness** dataset contains movie excerpts and key-frames and corresponding ground truth files based on classification into interesting and non-interesting samples. It is used for multimedia content interestingness classification. The dataset is composed of: - Shots and key-frames from a set of 78 Holly...
Provide a detailed description of the following dataset: Interestingness
AKCES-GEC
AKCES-GEC is a new dataset on grammatical error correction for Czech.
Provide a detailed description of the following dataset: AKCES-GEC
LASIESTA
**LASIESTA (Labeled and Annotated Sequences for Integral Evaluation of SegmenTation Algorithms)** is a segmentation and detection dataset composed by many real indoor and outdoor sequences organized into categories, each of one covering a specific challenge in moving object detection strategies.
Provide a detailed description of the following dataset: LASIESTA
E-GMD
Expanded Groove MIDI dataset (E-GMD) is an automatic drum transcription (ADT) dataset that contains 444 hours of audio from 43 drum kits, making it an order of magnitude larger than similar datasets, and the first with human-performed velocity annotations.
Provide a detailed description of the following dataset: E-GMD
FIRE
**Fundus Image Registration Dataset (FIRE)** is a dataset consisting of 129 retinal images forming 134 image pairs. These image pairs are split into 3 different categories depending on their characteristics. The images were acquired with a Nidek AFC-210 fundus camera, which acquires images with a resolution of 2912x291...
Provide a detailed description of the following dataset: FIRE
MVB
MVB (Multi View Baggage) is a dataset for baggage ReID task which has some essential differences from person ReID. The features of MVB are three-fold. First, MVB is the first publicly released large-scale dataset that contains 4519 baggage identities and 22660 annotated baggage images as well as its surface material la...
Provide a detailed description of the following dataset: MVB
300-VW
**300 Videos in the Wild (300-VW)** is a dataset for evaluating facial landmark tracking algorithms in the wild. The dataset authors collected a large number of long facial videos recorded in the wild. Each video has duration of ~1 minute (at 25-30 fps). All frames have been annotated with regards to the same mark-up (...
Provide a detailed description of the following dataset: 300-VW
Imp1k
Imp1k is a new dataset of designs annotated with importance information.
Provide a detailed description of the following dataset: Imp1k
PIROPO
The **PIROPO database** (People in Indoor ROoms with Perspective and Omnidirectional cameras) comprises multiple sequences recorded in two different indoor rooms, using both omnidirectional and perspective cameras. The sequences contain people in a variety of situations, including people walking, standing, and sitting....
Provide a detailed description of the following dataset: PIROPO
Million-AID
Million-AID is a large-scale benchmark dataset containing a million instances for RS scene classification. There are 51 semantic scene categories in Million-AID. And the scene categories are customized to match the land-use classification standards, which greatly enhance the practicability of the constructed Million-AI...
Provide a detailed description of the following dataset: Million-AID
Edge Milling Heads
The **Edge Milling Heads** data set comprises 144 images of an edge profile cutting head of a milling machine. The head tool contains a total of 30 cutting inserts. The cutting head is formed by 6 diagonals of inserts in radial direction along the tool perimeter, encompassing 5 inserts per diagonal in axial direction. ...
Provide a detailed description of the following dataset: Edge Milling Heads
MLe2e
MLe2 is a dataset for the evaluation of scene text end-to-end reading systems and all intermediate stages such as text detection, script identification and text recognition. The dataset contains a total of 711 scene images covering four different scripts (Latin, Chinese, Kannada, and Hangul).
Provide a detailed description of the following dataset: MLe2e
VxC TSG
The **VXC TSG** is based on samples taken from the ceramic tile industry and is comprised of 14 ceramic tile models, 42 surface grades and 960 pieces. It has been built in the VxC laboratory, at the Polytechnic University of Valencia, in collaboration with Keraben S.A., a large ceramic tile company located at Nules pro...
Provide a detailed description of the following dataset: VxC TSG
DADA-2000
DADA-2000 is a large-scale benchmark with 2000 video sequences (named as DADA-2000) is contributed with laborious annotation for driver attention (fixation, saccade, focusing time), accident objects/intervals, as well as the accident categories, and superior performance to state-of-the-arts are provided by thorough eva...
Provide a detailed description of the following dataset: DADA-2000
Panoramic Image Database
The **Panoramic Image Database** is a panoramic image dataset. The databases were collected by Andrew Vardy while visiting with the Computer Engineering group in February and March of 2004. Images were captured by a robot-mounted camera, pointed upwards at a hyperbolic mirror. The camera was an ImagingSource DFK 4303. ...
Provide a detailed description of the following dataset: Panoramic Image Database
SESIV
SEmantic Salient Instance Video (SESIV) dataset is obtained by augmenting the DAVIS-2017 benchmark dataset by assigning semantic ground-truth for salient instance labels. The SESIV dataset consists of 84 high-quality video sequences with pixel-wisely per-frame ground-truth labels.
Provide a detailed description of the following dataset: SESIV