dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
OTCBVS
**OCTCBVS** is a benchmark dataset for testing and evaluating novel and state-of-the-art computer vision algorithms. The benchmark contains videos and images recorded in and beyond the visible spectrum and is available for free to all researchers in the international computer vision communities.
Provide a detailed description of the following dataset: OTCBVS
LEAF-QA
LEAF-QA, a comprehensive dataset of 250,000 densely annotated figures/charts, constructed from real-world open data sources, along with ~2 million question-answer (QA) pairs querying the structure and semantics of these charts. LEAF-QA highlights the problem of multimodal QA, which is notably different from conventiona...
Provide a detailed description of the following dataset: LEAF-QA
Multi Task Crowd
Multi Task Crowd is a new 100 image dataset fully annotated for crowd counting, violent behaviour detection and density level classification.
Provide a detailed description of the following dataset: Multi Task Crowd
DogCentric Activity
The **DogCentric Activity** dataset is composed of dog activity videos taken from a first-person animal viewpoint. The dataset contains 10 different types of activities, including activities performed by the dog himself/herself, interactions between people and the dog, and activities performed by people or cars. Th...
Provide a detailed description of the following dataset: DogCentric Activity
Visual Question Answering v2.0
Visual Question Answering (VQA) v2.0 is a dataset containing open-ended questions about images. These questions require an understanding of vision, language and commonsense knowledge to answer. It is the second version of the [VQA](https://www.paperswithcode.com/dataset/vqa) dataset. - 265,016 images (COCO and abstr...
Provide a detailed description of the following dataset: Visual Question Answering v2.0
Biwi Kinect Head Pose
Biwi Kinect Head Pose is a challenging dataset mainly inspired by the automotive setup. It is acquired with the Microsoft Kinect sensor, a structured IR light device. It contains about 15k frame, with RGB. (640 × 480) and depth maps (640 × 480). Twenty subjects have been involved in the recordings: four of them were r...
Provide a detailed description of the following dataset: Biwi Kinect Head Pose
ELAS
ELAS is a dataset for lane detection. It contains more than 20 different scenes (in more than 15,000 frames) and considers a variety of scenarios (urban road, highways, traffic, shadows, etc.). The dataset was manually annotated for several events that are of interest for the research community (i.e., lane estimation, ...
Provide a detailed description of the following dataset: ELAS
100DOH
The 100 Days Of Hands Dataset (100DOH) is a large-scale video dataset containing hands and hand-object interactions. It consists of 27.3K Youtube videos from 11 categories with nearly 131 days of footage of everyday interaction. The focus of the dataset is hand contact, and it includes both first-person and third-perso...
Provide a detailed description of the following dataset: 100DOH
SVLD
The social vision and language dataset is a large-scale multimodal dataset designed for research into social contextual learning.
Provide a detailed description of the following dataset: SVLD
TextComplexityDE
TextComplexityDE is a dataset consisting of 1000 sentences in German language taken from 23 Wikipedia articles in 3 different article-genres to be used for developing text-complexity predictor models and automatic text simplification in German language. The dataset includes subjective assessment of different text-compl...
Provide a detailed description of the following dataset: TextComplexityDE
Image Paragraph Captioning
The Image Paragraph Captioning dataset allows researchers to benchmark their progress in generating paragraphs that tell a story about an image. The dataset contains 19,561 images from the [Visual Genome dataset](https://paperswithcode.com/dataset/visual-genome). Each image contains one paragraph. The training/val/test...
Provide a detailed description of the following dataset: Image Paragraph Captioning
Famulus
This is a dataset for segmentation and classification of epistemic activities in diagnostic reasoning texts.
Provide a detailed description of the following dataset: Famulus
CMU Wilderness Multilingual Speech Dataset
The CMU Wilderness Multilingual Speech Dataset is a dataset of over 700 different languages providing audio, aligned text and word pronunciations. On average each language provides around 20 hours of sentence-lengthed transcriptions.
Provide a detailed description of the following dataset: CMU Wilderness Multilingual Speech Dataset
Aesthetic Visual Analysis
**Aesthetic Visual Analysis** is a dataset for aesthetic image assessment that contains over 250,000 images along with a rich variety of meta-data including a large number of aesthetic scores for each image, semantic labels for over 60 categories as well as labels related to photographic style.
Provide a detailed description of the following dataset: Aesthetic Visual Analysis
BigBIRD
BigBIRD is a 3D dataset of 125 objects, with the following data for each object: * 600 12 megapixel images, sampling the viewing hemisphere * 600 registered RGB-D point clouds from a Carmine 1.09 sensor * Pose information for each of the above images and point clouds * Segmentation masks for each of the above ima...
Provide a detailed description of the following dataset: BigBIRD
WSJ0-2mix
**WSJ0-2mix** is a speech recognition corpus of speech mixtures using utterances from the Wall Street Journal (WSJ0) corpus.
Provide a detailed description of the following dataset: WSJ0-2mix
WHAM!
The **WSJ0 Hipster Ambient Mixtures** (**WHAM!**) dataset pairs each two-speaker mixture in the wsj0-2mix dataset with a unique noise background scene. It has an extension called [WHAMR!](/dataset/whamr) that adds artificial reverberation to the speech signals in addition to the background noise. The noise audio was...
Provide a detailed description of the following dataset: WHAM!
CUHK Face Alignment Database
The CUHK Face Alignment Database is dataset with 13,466 face images, among which 5, 590 images are from LFW and the remaining 7, 876 images are downloaded from the web. Each face is labeled with the positions of five facial keypoints. 10,000 images are used for training and the remaining 3,466 images for validation. ...
Provide a detailed description of the following dataset: CUHK Face Alignment Database
CUHK Square Dataset
CUHK Square data set is for transfer learning research on adapting generic pedestrian detectors. It includes a traffic video sequence of 60 minutes long. It is recorded by a stationary camera. The size of the scene is 720 by 576. In order to evaluate the performance of human detection on this data set, ground truth...
Provide a detailed description of the following dataset: CUHK Square Dataset
CUHK Occlusion Dataset
CUHK occlusion dataset includes 1,063 images with occluded pedestrians. It is used for Human Detection with occlusion handling in crowded scenes. Paper: [A discriminative deep model for pedestrian detection with occlusion handling](https://doi.org/10.1109/CVPR.2012.6248062)
Provide a detailed description of the following dataset: CUHK Occlusion Dataset
Grand Central Station Dataset
The Grand central station dataset includes a video with 50,010 frames which is used for Scene Understanding and Crowd Analysis. Paper: [Understanding collective crowd behaviors: Learning a Mixture model of Dynamic pedestrian-Agents](https://doi.org/10.1109/CVPR.2012.6248013)
Provide a detailed description of the following dataset: Grand Central Station Dataset
CUHK02
CUHK02 is a dataset for person re-identification. It contains 1,816 identities from two disjoint camera views. Each identity has two samples per camera view making a total of 7,264 images. It is used for Person Re-identification. Image Source: [Locally Aligned Feature Transforms across Views](https://www.cv-foundati...
Provide a detailed description of the following dataset: CUHK02
ArtEmis
ArtEmis is a large-scale dataset aimed at providing a detailed understanding of the interplay between visual content, its emotional effect, and explanations for the latter in language. In contrast to most existing annotation datasets in computer vision, this dataset focuses on the affective experience triggered by visu...
Provide a detailed description of the following dataset: ArtEmis
BreakHis
The Breast Cancer Histopathological Image Classification (BreakHis) is composed of 9,109 microscopic images of breast tumor tissue collected from 82 patients using different magnifying factors (40X, 100X, 200X, and 400X). It contains 2,480 benign and 5,429 malignant samples (700X460 pixels, 3-channel RGB, 8-bit dept...
Provide a detailed description of the following dataset: BreakHis
2D Hela
2D HeLa is a dataset of fluorescence microscopy images of HeLa cells stained with various organelle-specific fluorescent dyes. The images include 10 organelles, which are DNA (Nuclei), ER (Endoplasmic reticulum), Giantin, (cis/medial Golgi), GPP130 (cis Golgi), Lamp2 (Lysosomes), Mitochondria, Nucleolin (Nucleoli), Act...
Provide a detailed description of the following dataset: 2D Hela
PointPattern
PointPattern is a graph classification dataset constructed by simple point patterns from statistical mechanics. The authors simulated three point patterns in 2D: hard disks in equilibrium (HD), Poisson point process, and random sequential adsorption (RSA) of disks. The HD and Poisson distributions can be seen as simple...
Provide a detailed description of the following dataset: PointPattern
Humans in 3D
H3D (Humans in 3D) is a dataset of annotated people. The annotations include: * The joints and other keypoints (eyes, ears, nose, shoulders, elbows, wrists, hips, knees and ankles) * The 3D pose inferred from the keypoints. * Visibility boolean for each keypoint * Region annotations (upper clothes, lower clothes,...
Provide a detailed description of the following dataset: Humans in 3D
BelgaLogos
BelgaLogos is a dataset for logo detection and recognition. The images of BelgaLogos dataset have been provided and are copyrighted by BELGA press agency. They are freely available for research purpose only. The dataset is composed of 10,000 images covering all aspects of life and current affairs: politics and economic...
Provide a detailed description of the following dataset: BelgaLogos
Aspects dataset
This dataset contains video shots for two different classes: tigers and cars. The shots were collected from 188 car ads (~1–2 min each) and 14 nature documentaries about tigers (~40 min), amounting to roughly 14 h of video. The videos were partitioned into shorter shots, and only those showing at least one instance of ...
Provide a detailed description of the following dataset: Aspects dataset
POET
The POET (Pascal Objects Eye Tracking) is a dataset that consists of eye tracking data for the complete trainval set of ten objects classes (cat, dog, bicycle, motorbike, boat, aeroplane, horse, cow, sofa, dining table) from [Pascal VOC 2012](pascal-voc) (6,270 images in total). Each image is annotated with the eye mov...
Provide a detailed description of the following dataset: POET
AMUSE
The automotive multi-sensor (AMUSE) dataset consists of inertial and other complementary sensor data combined with monocular, omnidirectional, high frame rate visual data taken in real traffic scenes during multiple test drives. Paper: [A Multi-sensor Traffic Scene Dataset with Omnidirectional Video](https://doi.org...
Provide a detailed description of the following dataset: AMUSE
IMO
Dataset of annotated independently moving objects (IMO). This dataset contains left and right images, stereo images, stereo disparity from SGM, and vehicle labels as well as a ground truth annotations. Paper: [Independently Moving Object Trajectories from Sequential Hierarchical Ransac](https://users.isy.liu.se/cvl...
Provide a detailed description of the following dataset: IMO
LTIR
The LTIR dataset is a thermal infrared dataset for evaluation of Short-Term Single-Object (STSO) tracking. The dataset contains * 20 thermal infrared sequences, one .png per frame. Some sequences are available in both 8- and 16-bits. * Bounding box annotations of one object per sequence. * Local per-frame annot...
Provide a detailed description of the following dataset: LTIR
Family101
The Family101 dataset is the a large-scale dataset of families across several generations. It contains 101 different families with distinct family names, including 206 nuclear families, 607 individuals, with 14,816 images. The dataset are composed of renowned public families. Paper: [Kinship Classification by Modeli...
Provide a detailed description of the following dataset: Family101
FIW
FIW is a large and comprehensive database available for kinship recognition. FIW is made up of 11,932 natural family photos of 1,000 families-- nearly 10x more than the next-to-largest, [Family-101](family101) database. Also, it contains 656,954 image pairs split between the 11 relationships, which is much larger than ...
Provide a detailed description of the following dataset: FIW
KinFaceW
KinFaceW consists of two kinship datasets: KinFaceW-I and KinFaceW-II. Face images were collected from the internet, including some public figure face images as well as their parents' or children's face images. Face images are captured under uncontrolled environments in two datasets with no restriction in terms of pose...
Provide a detailed description of the following dataset: KinFaceW
Boxy
A large vehicle detection dataset with almost two million annotated vehicles for training and evaluating object detection methods for self-driving cars on freeways. The dataset consists of: * 200,000 images * 1,990,000 annotated vehicles * 5 Megapixel resolution * Sunshine, rain, dusk, night * Clear freeways,...
Provide a detailed description of the following dataset: Boxy
CASR
CASR is a dataset for cyclist arm signal recognition in videos. It contains 219 annotated arm signal actions on videos of approximately 10 seconds each, containing one or two actions per video.
Provide a detailed description of the following dataset: CASR
Driving Event Camera Dataset
This dataset consists of a number of sequences that were recorded with a VGA (640x480) event camera (Samsung DVS Gen3) and a conventional RGB camera (Huawei P20 Pro) placed on the windshield of a car driving through Zurich.
Provide a detailed description of the following dataset: Driving Event Camera Dataset
FRIDA
FRIDA and FRIDA2 are databases of numerical images easily usable to evaluate in a systematic way the performance of visibility and contrast restoration algorithms. FRIDA comprises 90 synthetic images of 18 urban road scenes. FRIDA2 comprises 330 synthetic images of 66 diverse road scenes. The view point is closed to th...
Provide a detailed description of the following dataset: FRIDA
Ford Campus Vision and Lidar Data Set
Ford Campus Vision and Lidar Data Set is a dataset collected by an autonomous ground vehicle testbed, based upon a modified Ford F-250 pickup truck. The vehicle is outfitted with a professional (Applanix POS LV) and consumer (Xsens MTI-G) Inertial Measuring Unit (IMU), a Velodyne 3D-lidar scanner, two push-broom forwar...
Provide a detailed description of the following dataset: Ford Campus Vision and Lidar Data Set
JAAD
JAAD is a dataset for studying joint attention in the context of autonomous driving. The focus is on pedestrian and driver behaviors at the point of crossing and factors that influence them. To this end, JAAD dataset provides a richly annotated collection of 346 short video clips (5-10 sec long) extracted from over 240...
Provide a detailed description of the following dataset: JAAD
LISA Vehicle Detection
This is a dataset for vehicle detection. It consists of: * Three color video sequences captured at different times of the day and illumination settings: morning, evening, sunny, cloudy, etc. * Different driving environments: highway and urban. * Varying traffic conditions: light to dense traffic Paper: [A Gener...
Provide a detailed description of the following dataset: LISA Vehicle Detection
LLAMAS
The unsupervised Labeled Lane MArkerS dataset (LLAMAS) is a dataset for lane detection and segmentation. It contains over 100,000 annotated images, with annotations of over 100 meters at a resolution of 1276 x 717 pixels. The Unsupervised Llamas dataset was annotated by creating high definition maps for automated drivi...
Provide a detailed description of the following dataset: LLAMAS
VIsual PERception (VIPER)
VIPER is a benchmark suite for visual perception. The benchmark is based on more than 250K high-resolution video frames, all annotated with ground-truth data for both low-level and high-level vision tasks, including optical flow, semantic instance segmentation, object detection and tracking, object-level 3D scene layou...
Provide a detailed description of the following dataset: VIsual PERception (VIPER)
REC-COCO
Relations in Captions (REC-COCO) is a new dataset that contains associations between caption tokens and bounding boxes in images. REC-COCO is based on the MS-COCO and V-COCO datasets. For each image in V-COCO, we collect their corresponding captions from MS-COCO and automatically align the concept triplet in V-COCO to ...
Provide a detailed description of the following dataset: REC-COCO
TRIPOD
TRIPOD contains screenplays and plot synopses with turning point (TP) annotations for 99 movies. Each movie contains: 1. The Wikipedia plot synopsis (extended summary of 35 sentences on average) with sentence-level TP annotations. 2. The screenplay (all dialogue and description parts of the movie) segmented into sc...
Provide a detailed description of the following dataset: TRIPOD
CSI Screenplay Summarization Corpus
The dataset contains gold-standard summary labels for 39 "CSI: Crime Scene Investigation" episodes from seasons 1-5. Each episode contains the full-length screenplay and human annotations for its summary. The annotations include: 1. scene-level binary labels denoting whether the scene belongs to the summary of the e...
Provide a detailed description of the following dataset: CSI Screenplay Summarization Corpus
FPV-O
FPV-O is a multi-subject first-person vision dataset of office activities. Office activities include person-to-person interactions, such as chatting and handshaking, person-to-object interactions, such as using a computer or a whiteboard, as well as generic activities such as walking. The videos in the dataset present ...
Provide a detailed description of the following dataset: FPV-O
MERL Shopping
MERL Shopping is a dataset for training and testing action detection algorithms. The MERL Shopping Dataset consists of 106 videos, each of which is a sequence about 2 minutes long. The videos are from a fixed overhead camera looking down at people shopping in a grocery store setting. Each video contains several instanc...
Provide a detailed description of the following dataset: MERL Shopping
A2D
A2D (Actor-Action Dataset) is a dataset for simultaneously inferring actors and actions in videos. A2D has seven actor classes (adult, baby, ball, bird, car, cat, and dog) and eight action classes (climb, crawl, eat, fly, jump, roll, run, and walk) not including the no-action class, which we also consider. The A2D has ...
Provide a detailed description of the following dataset: A2D
ASD
The Annotated Semantic Dataset is composed of $11$ videos, divided in $3$ activity categories: Biking; Driving and Walking, according to their amount of semantic information. The classes are: $0p$, which represents the videos with approximately no semantic information; $25p$, for the videos containing relevant semantic...
Provide a detailed description of the following dataset: ASD
l2d
This dataset is composed of paired videos of people dancing 3 different music styles: Ballet, Michael Jackson and Salsa. It contains multimodal data (visual data, temporal-graphs and audio) careful-selected from publicly available videos of dancers performing representative movements of the music style and audio data ...
Provide a detailed description of the following dataset: l2d
OccludedPASCAL3D+
The **OccludedPASCAL3D+** is a dataset is designed to evaluate the robustness to occlusion for a number of computer vision tasks, such as object detection, keypoint detection and pose estimation. In the OccludedPASCAL3D+ dataset, we simulate partial occlusion by superimposing objects cropped from the MS-COCO dataset on...
Provide a detailed description of the following dataset: OccludedPASCAL3D+
THEODORE
Recent work about synthetic indoor datasets from perspective views has shown significant improvements of object detection results with Convolutional Neural Networks(CNNs). In this paper, we introduce THEODORE: a novel, large-scale indoor dataset containing 100,000 high- resolution diversified fisheye images with 14 cla...
Provide a detailed description of the following dataset: THEODORE
MHRI dataset
The dataset includes recordings from 10 different users teaching the robot different common kitchen objects, that consists of synchronized recordings from three cameras and a microphone mounted on the robot: An RGB-d camera covers the user manipulation and interaction with the robot An RGB-d camera mounted ...
Provide a detailed description of the following dataset: MHRI dataset
highD Dataseth
The highD dataset is a new dataset of naturalistic vehicle trajectories recorded on German highways. Using a drone, typical limitations of established traffic data collection methods such as occlusions are overcome by the aerial perspective. Traffic was recorded at six different locations and includes more than 110 500...
Provide a detailed description of the following dataset: highD Dataseth
inD Dataset
The **inD** dataset is a new dataset of naturalistic vehicle trajectories recorded at German intersections. Using a drone, typical limitations of established traffic data collection methods like occlusions are overcome. Traffic was recorded at four different locations. The trajectory for each road user and its type is ...
Provide a detailed description of the following dataset: inD Dataset
rounD Dataset
The rounD dataset is a new dataset of naturalistic road user trajectories recorded at German roundabouts. Using a drone, typical limitations of established traffic data collection methods like occlusions are overcome. Traffic was recorded at three different locations. The trajectory for each road user and its type is e...
Provide a detailed description of the following dataset: rounD Dataset
Localized Narratives
We propose Localized Narratives, a new form of multimodal image annotations connecting vision and language. We ask annotators to describe an image with their voice while simultaneously hovering their mouse over the region they are describing. Since the voice and the mouse pointer are synchronized, we can localize every...
Provide a detailed description of the following dataset: Localized Narratives
CE4
Given the difficulty to handle planetary data we provide downloadable files in PNG format from the missions Chang'E-3 and Chang'E-4. In addition to a set of scripts to do the conversion given a different PDS4 Dataset. This set of images constitute one of the first available datasets to tackle problems of Computer V...
Provide a detailed description of the following dataset: CE4
MICC-SRI
The dataset contains 11,913 frame pairs of urban driving footage with and without moving objects, synthetically generated with the CARLA simulator. All frames are available both as RGB images and semantic segmentations. RGB images are non-photorealistic being rendered by a game engine, while semantic segmentations are ...
Provide a detailed description of the following dataset: MICC-SRI
KITTI-trajectory-prediction
KITTI is a well established dataset in the computer vision community. It has often been used for trajectory prediction despite not having a well defined split, generating non comparable baselines in different works. This dataset aims at bridging this gap and proposes a well defined split of the KITTI data. Samples are...
Provide a detailed description of the following dataset: KITTI-trajectory-prediction
EmoContext
EmoContext consists of three-turn English Tweets. The emotion labels include happiness, sadness, anger and other.
Provide a detailed description of the following dataset: EmoContext
Glint360K
The largest and cleanest face recognition dataset Glint360K, which contains **`17,091,657`** images of **`360,232`** individuals, baseline models trained on Glint360K can easily achieve state-of-the-art performance.
Provide a detailed description of the following dataset: Glint360K
IndicCorp
IndicCorp is a large monolingual corpora with around 9 billion tokens covering 12 of the major Indian languages. It has been developed by discovering and scraping thousands of web sources - primarily news, magazines and books, over a duration of several months. **Languages covered**: Assamese, Bengali, English, Guja...
Provide a detailed description of the following dataset: IndicCorp
RuFa
RuFa (Ruqaa-Farsi) dataset contains images of text written in one of two Arabic fonts: Ruqaa and Nastaliq (Farsi). The dataset contains 40,000 synthesized image and 516 real one, 40,516 in total. Images are in RGB JPG format at 100×100px. Text in the images has varying number of words, position, size, and opacity. R...
Provide a detailed description of the following dataset: RuFa
MERL-RAV
The MERL-RAV (MERL Reannotation of AFLW with Visibility) Dataset contains over 19,000 face images in a full range of head poses. Each face is manually labeled with the ground-truth locations of 68 landmarks, with the additional information of whether each landmark is unoccluded, self-occluded (due to extreme head poses...
Provide a detailed description of the following dataset: MERL-RAV
News Interactions on Globo.com
### Context This large dataset with users interactions logs (page views) from a news portal was kindly provided by [Globo.com][1], the most popular news portal in Brazil, for reproducibility of the experiments with CHAMELEON - a meta-architecture for contextual hybrid session-based news recommender systems. The sour...
Provide a detailed description of the following dataset: News Interactions on Globo.com
Synbols
Synbols is a dataset generator designed for probing the behavior of learning algorithms. By defining the distribution over latent factors one can craft a dataset specifically tailored to answer specific questions about a given algorithm. Default versions of these datasets are also materialized and can serve as bench...
Provide a detailed description of the following dataset: Synbols
C&Z
One of the first datasets (if not the first) to highlight the importance of bias and diversity in the community, which started a revolution afterwards. Introduced in 2014 as integral part of a thesis of Master of Science [1,2] at Carnegie Mellon and City University of Hong Kong. It was later expanded by adding syntheti...
Provide a detailed description of the following dataset: C&Z
GEM
Generation, Evaluation, and Metrics (GEM) is a benchmark environment for Natural Language Generation with a focus on its Evaluation, both through human annotations and automated Metrics. GEM aims to: - measure NLG progress across 13 datasets spanning many NLG tasks and languages. - provide an in-depth analysis o...
Provide a detailed description of the following dataset: GEM
ALFWorld
ALFWorld contains interactive TextWorld environments (Côté et. al) that parallel embodied worlds in the ALFRED dataset (Shridhar et. al). The aligned environments allow agents to reason and learn high-level policies in an abstract space before solving embodied tasks through low-level actuation.
Provide a detailed description of the following dataset: ALFWorld
HQ-WMCA
The High-Quality Wide Multi-Channel Attack database (HQ-WMCA) database consists of 2904 short multi-modal video recordings of both bona-fide and presentation attacks. There are 555 bonafide presentations from 51 participants and the remaining 2349 are presentation attacks. The data is recorded from several channels inc...
Provide a detailed description of the following dataset: HQ-WMCA
The Best Sarcasm Annotated Dataset in Spanish
### Content This dataset contains all utterances of two episodes of South Park (Latin American voices) and two episodes of Archer (Spanish voices). The order of the utterances is shuffled. Each utterance has been annotated based on whether it is sarcastic or not. Sarcastic expressions also contain further annotation...
Provide a detailed description of the following dataset: The Best Sarcasm Annotated Dataset in Spanish
MIRACL-VC1
MIRACL-VC1 is a lip-reading dataset including both depth and color images. It can be used for diverse research fields like visual speech recognition, face detection, and biometrics. Fifteen speakers (five men and ten women) positioned in the frustum of an MS Kinect sensor and utter ten times a set of ten words and ten ...
Provide a detailed description of the following dataset: MIRACL-VC1
XD-Violence
XD-Violence is a large-scale audio-visual dataset for violence detection in videos.
Provide a detailed description of the following dataset: XD-Violence
PatentMatch
We address the computer-assisted search for prior art by creating a training dataset for supervised machine learning called PatentMatch. It contains pairs of claims from patent applications and semantically corresponding text passages of different degrees from cited patent documents. Each pair has been labeled by techn...
Provide a detailed description of the following dataset: PatentMatch
A Dataset of Journalists' Interactions with Their Readership
We present a dataset of dialogs in which journalists of The Guardian replied to reader comments and identify the reasons why. Based on this data, we formulate the novel task of recommending reader comments to journalists that are worth reading or replying to, i.e., ranking comments in such a way that the top comments a...
Provide a detailed description of the following dataset: A Dataset of Journalists' Interactions with Their Readership
Top Comment or Flop Comment?
This dataset comprises four files of IDs of either strongly or weakly engaging online news comments (please see the paper for details): "Top comments" are 1) the top 10% comments in the politics section of The Guardian with the largest relative number of *replies* received (3111 samples) and 2) the top 10% comments in...
Provide a detailed description of the following dataset: Top Comment or Flop Comment?
HeartSeg
The medaka (Oryzias latipes) and the zebrafish (Danio rerio) are used as a model organism for a variety of subjects in biomedical research. The presented work aims to study the potential of automated ventricular dimension estimation through heart segmentation in medaka. For more on this, it's time for a closer look on ...
Provide a detailed description of the following dataset: HeartSeg
DNS Challenge
The DNS Challenge at INTERSPEECH 2020 intended to promote collaborative research in single-channel Speech Enhancement aimed to maximize the perceptual quality and intelligibility of the enhanced speech. The challenge evaluated the speech quality using the online subjective evaluation framework ITU-T P.808. The challeng...
Provide a detailed description of the following dataset: DNS Challenge
Interspeech 2021 Deep Noise Suppression Challenge
The Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality. This challenge has two two tracks: **Track 1: Real-Time Denoising track for wide band scenario** The noise suppressor must take less than the stride time T...
Provide a detailed description of the following dataset: Interspeech 2021 Deep Noise Suppression Challenge
TRN
The Toulouse Road Network dataset describes patches of road maps from the city of Toulouse, represented both as spatial graphs G = (A, X) and as grayscale segmentation images. The TRN dataset contains 111,034 data points (map tiles), of which: 80,357 are in the training set (around 72.4%), 11,679 are in the validat...
Provide a detailed description of the following dataset: TRN
WEB-FORUM-52
The WEB-FORUM-52 gold standard comprises (i) 13 web forums from the health domain, (ii) 15 forums obtained from a Wikipedia list of popular forums (https://en.wikipedia.org/wiki/List_of_Internet_forums), (iii) 13 forums mentioned on a list of popular German Web forums (https://www.beliebte-foren.de), (iv) nine forums o...
Provide a detailed description of the following dataset: WEB-FORUM-52
KorQuAD
KorQuAD is a large-scale question-and-answer dataset constructed for Korean machine reading comprehension, and investigate the dataset to understand the distribution of answers and the types of reasoning required to answer the question. This dataset benchmarks the data generating process of SQuAD to meet the standard.
Provide a detailed description of the following dataset: KorQuAD
MOBIO
The MOBIO database consists of bi-modal (audio and video) data taken from 152 people. The database has a female-male ratio or nearly 1:2 (100 males and 52 females) and was collected from August 2008 until July 2010 in six different sites from five different countries. This led to a diverse bi-modal database with both n...
Provide a detailed description of the following dataset: MOBIO
FRLL-Morphs
FRLL-Morphs is a dataset of morphed faces based on images selected from the publicly available Face Research London Lab dataset [1]. We created the database by selecting similar looking pairs of people, and made 4 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGA...
Provide a detailed description of the following dataset: FRLL-Morphs
VisualMRC
VisualMRC is a visual machine reading comprehension dataset that proposes a task: given a question and a document image, a model produces an abstractive answer. You can find more details, analyses, and baseline results in the paper, VisualMRC: Machine Reading Comprehension on Document Images, AAAI 2021. Stati...
Provide a detailed description of the following dataset: VisualMRC
FERET-Morphs
FERET-Morphs is a dataset of morphed faces selected from the publicly available FERET dataset [1]. We created the database by selecting similar looking pairs of people, and made 3 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGAN 2 [3]. * [1] https://www.nist....
Provide a detailed description of the following dataset: FERET-Morphs
FRGC-Morphs
FRGC-Morphs is a dataset of morphed faces selected from the publicly available FRGC dataset [1]. We created the database by selecting similar looking pairs of people, and made 3 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGAN 2 [3]. * [1] https://www.nist.go...
Provide a detailed description of the following dataset: FRGC-Morphs
NISP- A Multi-lingual Multi-accent Dataset for Speaker Profiling
We announce the release of a new multilingual speaker dataset called NITK-IISc Multilingual Multi-accent Speaker Profiling(NISP) dataset. The dataset contains speech in six different languages -- five Indian languages along with Indian English. The dataset contains speech data from 345 bilingual speakers in India. Each...
Provide a detailed description of the following dataset: NISP- A Multi-lingual Multi-accent Dataset for Speaker Profiling
NinaPro DB2
The second Ninapro database includes 40 intact subjects and it is thoroughly described in the paper: "Manfredo Atzori, Arjan Gijsberts, Claudio Castellini, Barbara Caputo, Anne-Gabrielle Mittaz Hager, Simone Elsig, Giorgio Giatsidis, Franco Bassetto & Henning Müller. Electromyography data for non-invasive naturally-con...
Provide a detailed description of the following dataset: NinaPro DB2
BuzzFeed-Webis Fake News Corpus 2016
The BuzzFeed-Webis Fake News Corpus 16 comprises the output of 9 publishers in a week close to the US elections. Among the selected publishers are 6 prolific hyperpartisan ones (three left-wing and three right-wing), and three mainstream publishers (see Table 1). All publishers earned Facebook’s blue checkmark, indicat...
Provide a detailed description of the following dataset: BuzzFeed-Webis Fake News Corpus 2016
POLIT-FALSE-n-LEGIT NEWS DB 2016-2017
The LiT.RL POLIT-FALSE-n-LEGIT NEWS DB 2016-2017 contains a total of 274 news articles about U.S. Politics, content-matched in pairs of legitimate and falsified news. The database is free and released under an open license for educational and research purposes.
Provide a detailed description of the following dataset: POLIT-FALSE-n-LEGIT NEWS DB 2016-2017
GQN rooms-ring-camera
GQN rooms-ring-camera consist of scenes of a variable number of random objects captured in a square room of size 7x7 units. Wall textures, floor textures as well as the shapes of the objects are randomly chosen within a fixed pool of discrete options. There are 5 possible wall textures (red, green, cerise, orange, yell...
Provide a detailed description of the following dataset: GQN rooms-ring-camera
ISOT Fake News Dataset
The ISOT Fake News dataset is a compilation of several thousands fake news and truthful articles, obtained from different legitimate news sites and sites flagged as unreliable by Politifact.com.
Provide a detailed description of the following dataset: ISOT Fake News Dataset
ObjectsRoom
The **ObjectsRoom** dataset is based on the MuJoCo environment used by the Generative Query Network [4] and is a multi-object extension of the 3d-shapes dataset. The training set contains 1M scenes with up to three objects. We also provide ~1K test examples for the following variants: 2.1 Empty room: scenes consist ...
Provide a detailed description of the following dataset: ObjectsRoom
SVDC Fake News Dataset
A labeled dataset that presents fake news surrounding the conflict in Syria. The dataset consists of a set of articles/news labeled by 0 (fake) or 1 (credible). Credibility of articles are computed with respect to a ground truth information obtained from the Syrian Violations Documentation Center (VDC). In particular,...
Provide a detailed description of the following dataset: SVDC Fake News Dataset
FakeNewsAMT & Celebrity
**FakeNewsAMT & Celebrity** include two novel datasets for the task of fake news detection, covering seven different news domains.
Provide a detailed description of the following dataset: FakeNewsAMT & Celebrity