dataset_name stringlengths 2 128 | description stringlengths 1 9.7k | prompt stringlengths 59 185 |
|---|---|---|
OTCBVS | **OCTCBVS** is a benchmark dataset for testing and evaluating novel and state-of-the-art computer vision algorithms. The benchmark contains videos and images recorded in and beyond the visible spectrum and is available for free to all researchers in the international computer vision communities. | Provide a detailed description of the following dataset: OTCBVS |
LEAF-QA | LEAF-QA, a comprehensive dataset of 250,000 densely annotated figures/charts, constructed from real-world open data sources, along with ~2 million question-answer (QA) pairs querying the structure and semantics of these charts. LEAF-QA highlights the problem of multimodal QA, which is notably different from conventiona... | Provide a detailed description of the following dataset: LEAF-QA |
Multi Task Crowd | Multi Task Crowd is a new 100 image dataset fully annotated for crowd counting, violent behaviour detection and density level classification. | Provide a detailed description of the following dataset: Multi Task Crowd |
DogCentric Activity | The **DogCentric Activity** dataset is composed of dog activity videos taken from a first-person animal viewpoint. The dataset contains 10 different types of activities, including activities performed by the dog himself/herself, interactions between people and the dog, and activities performed by people or cars.
Th... | Provide a detailed description of the following dataset: DogCentric Activity |
Visual Question Answering v2.0 | Visual Question Answering (VQA) v2.0 is a dataset containing open-ended questions about images. These questions require an understanding of vision, language and commonsense knowledge to answer. It is the second version of the [VQA](https://www.paperswithcode.com/dataset/vqa) dataset.
- 265,016 images (COCO and abstr... | Provide a detailed description of the following dataset: Visual Question Answering v2.0 |
Biwi Kinect Head Pose | Biwi Kinect Head Pose is a challenging dataset mainly inspired by the automotive setup. It is acquired with the Microsoft Kinect sensor, a structured IR light device. It contains about 15k frame, with RGB. (640 × 480) and depth maps (640 × 480). Twenty subjects have been involved in the recordings: four of them were r... | Provide a detailed description of the following dataset: Biwi Kinect Head Pose |
ELAS | ELAS is a dataset for lane detection. It contains more than 20 different scenes (in more than 15,000 frames) and considers a variety of scenarios (urban road, highways, traffic, shadows, etc.). The dataset was manually annotated for several events that are of interest for the research community (i.e., lane estimation, ... | Provide a detailed description of the following dataset: ELAS |
100DOH | The 100 Days Of Hands Dataset (100DOH) is a large-scale video dataset containing hands and hand-object interactions. It consists of 27.3K Youtube videos from 11 categories with nearly 131 days of footage of everyday interaction. The focus of the dataset is hand contact, and it includes both first-person and third-perso... | Provide a detailed description of the following dataset: 100DOH |
SVLD | The social vision and language dataset is a large-scale multimodal dataset designed for research into social contextual learning. | Provide a detailed description of the following dataset: SVLD |
TextComplexityDE | TextComplexityDE is a dataset consisting of 1000 sentences in German language taken from 23 Wikipedia articles in 3 different article-genres to be used for developing text-complexity predictor models and automatic text simplification in German language. The dataset includes subjective assessment of different text-compl... | Provide a detailed description of the following dataset: TextComplexityDE |
Image Paragraph Captioning | The Image Paragraph Captioning dataset allows researchers to benchmark their progress in generating paragraphs that tell a story about an image. The dataset contains 19,561 images from the [Visual Genome dataset](https://paperswithcode.com/dataset/visual-genome). Each image contains one paragraph. The training/val/test... | Provide a detailed description of the following dataset: Image Paragraph Captioning |
Famulus | This is a dataset for segmentation and classification of epistemic activities in diagnostic reasoning texts. | Provide a detailed description of the following dataset: Famulus |
CMU Wilderness Multilingual Speech Dataset | The CMU Wilderness Multilingual Speech Dataset is a dataset of over 700 different languages providing audio, aligned text and word pronunciations. On average each language provides around 20 hours of sentence-lengthed transcriptions. | Provide a detailed description of the following dataset: CMU Wilderness Multilingual Speech Dataset |
Aesthetic Visual Analysis | **Aesthetic Visual Analysis** is a dataset for aesthetic image assessment that contains over 250,000 images along with a rich variety of meta-data including a
large number of aesthetic scores for each image, semantic labels for over 60 categories as well as labels related to photographic style. | Provide a detailed description of the following dataset: Aesthetic Visual Analysis |
BigBIRD | BigBIRD is a 3D dataset of 125 objects, with the following data for each object:
* 600 12 megapixel images, sampling the viewing hemisphere
* 600 registered RGB-D point clouds from a Carmine 1.09 sensor
* Pose information for each of the above images and point clouds
* Segmentation masks for each of the above ima... | Provide a detailed description of the following dataset: BigBIRD |
WSJ0-2mix | **WSJ0-2mix** is a speech recognition corpus of speech mixtures using utterances from the Wall Street Journal (WSJ0) corpus. | Provide a detailed description of the following dataset: WSJ0-2mix |
WHAM! | The **WSJ0 Hipster Ambient Mixtures** (**WHAM!**) dataset pairs each two-speaker mixture in the wsj0-2mix dataset with a unique noise background scene. It has an extension called [WHAMR!](/dataset/whamr) that adds artificial reverberation to the speech signals in addition to the background noise.
The noise audio was... | Provide a detailed description of the following dataset: WHAM! |
CUHK Face Alignment Database | The CUHK Face Alignment Database is dataset with 13,466 face images, among which 5, 590 images are from LFW and the remaining 7, 876 images are downloaded from the web. Each face is labeled with the positions of five facial keypoints. 10,000 images are used for training and the remaining 3,466 images for validation.
... | Provide a detailed description of the following dataset: CUHK Face Alignment Database |
CUHK Square Dataset | CUHK Square data set is for transfer learning research on adapting generic pedestrian detectors. It includes a traffic video sequence of 60 minutes long. It is recorded by a stationary camera. The size of the scene is 720 by 576.
In order to evaluate the performance of human detection on this data set, ground truth... | Provide a detailed description of the following dataset: CUHK Square Dataset |
CUHK Occlusion Dataset | CUHK occlusion dataset includes 1,063 images with occluded pedestrians. It is used for Human Detection with occlusion handling in crowded scenes.
Paper: [A discriminative deep model for pedestrian detection with occlusion handling](https://doi.org/10.1109/CVPR.2012.6248062) | Provide a detailed description of the following dataset: CUHK Occlusion Dataset |
Grand Central Station Dataset | The Grand central station dataset includes a video with 50,010 frames which is used for Scene Understanding and Crowd Analysis.
Paper: [Understanding collective crowd behaviors: Learning a Mixture model of Dynamic pedestrian-Agents](https://doi.org/10.1109/CVPR.2012.6248013) | Provide a detailed description of the following dataset: Grand Central Station Dataset |
CUHK02 | CUHK02 is a dataset for person re-identification. It contains 1,816 identities from two disjoint camera views. Each identity has two samples per camera view making a total of 7,264 images. It is used for Person Re-identification.
Image Source: [Locally Aligned Feature Transforms across Views](https://www.cv-foundati... | Provide a detailed description of the following dataset: CUHK02 |
ArtEmis | ArtEmis is a large-scale dataset aimed at providing a detailed understanding of the interplay between visual content, its emotional effect, and explanations for the latter in language. In contrast to most existing annotation datasets in computer vision, this dataset focuses on the affective experience triggered by visu... | Provide a detailed description of the following dataset: ArtEmis |
BreakHis | The Breast Cancer Histopathological Image Classification (BreakHis) is composed of 9,109 microscopic images of breast tumor tissue collected from 82 patients using different magnifying factors (40X, 100X, 200X, and 400X). It contains 2,480 benign and 5,429 malignant samples (700X460 pixels, 3-channel RGB, 8-bit dept... | Provide a detailed description of the following dataset: BreakHis |
2D Hela | 2D HeLa is a dataset of fluorescence microscopy images of HeLa cells stained with various organelle-specific fluorescent dyes. The images include 10 organelles, which are DNA (Nuclei), ER (Endoplasmic reticulum), Giantin, (cis/medial Golgi), GPP130 (cis Golgi), Lamp2 (Lysosomes), Mitochondria, Nucleolin (Nucleoli), Act... | Provide a detailed description of the following dataset: 2D Hela |
PointPattern | PointPattern is a graph classification dataset constructed by simple point patterns from statistical mechanics. The authors simulated three point patterns in 2D: hard disks in equilibrium (HD), Poisson point process, and random sequential adsorption (RSA) of disks. The HD and Poisson distributions can be seen as simple... | Provide a detailed description of the following dataset: PointPattern |
Humans in 3D | H3D (Humans in 3D) is a dataset of annotated people. The annotations include:
* The joints and other keypoints (eyes, ears, nose, shoulders, elbows, wrists, hips, knees and ankles)
* The 3D pose inferred from the keypoints.
* Visibility boolean for each keypoint
* Region annotations (upper clothes, lower clothes,... | Provide a detailed description of the following dataset: Humans in 3D |
BelgaLogos | BelgaLogos is a dataset for logo detection and recognition. The images of BelgaLogos dataset have been provided and are copyrighted by BELGA press agency. They are freely available for research purpose only. The dataset is composed of 10,000 images covering all aspects of life and current affairs: politics and economic... | Provide a detailed description of the following dataset: BelgaLogos |
Aspects dataset | This dataset contains video shots for two different classes: tigers and cars. The shots were collected from 188 car ads (~1–2 min each) and 14 nature documentaries about tigers (~40 min), amounting to roughly 14 h of video. The videos were partitioned into shorter shots, and only those showing at least one instance of ... | Provide a detailed description of the following dataset: Aspects dataset |
POET | The POET (Pascal Objects Eye Tracking) is a dataset that consists of eye tracking data for the complete trainval set of ten objects classes (cat, dog, bicycle, motorbike, boat, aeroplane, horse, cow, sofa, dining table) from [Pascal VOC 2012](pascal-voc) (6,270 images in total). Each image is annotated with the eye mov... | Provide a detailed description of the following dataset: POET |
AMUSE | The automotive multi-sensor (AMUSE) dataset consists of inertial and other complementary sensor data combined with monocular, omnidirectional, high frame rate visual data taken in real traffic scenes during multiple test drives.
Paper: [A Multi-sensor Traffic Scene Dataset with Omnidirectional Video](https://doi.org... | Provide a detailed description of the following dataset: AMUSE |
IMO | Dataset of annotated independently moving objects (IMO). This dataset contains left and right images, stereo images, stereo disparity from SGM, and vehicle labels as well as a ground truth annotations.
Paper: [Independently Moving Object Trajectories from Sequential Hierarchical Ransac](https://users.isy.liu.se/cvl... | Provide a detailed description of the following dataset: IMO |
LTIR | The LTIR dataset is a thermal infrared dataset for evaluation of Short-Term Single-Object (STSO) tracking.
The dataset contains
* 20 thermal infrared sequences, one .png per frame. Some sequences are available in both 8- and 16-bits.
* Bounding box annotations of one object per sequence.
* Local per-frame annot... | Provide a detailed description of the following dataset: LTIR |
Family101 | The Family101 dataset is the a large-scale dataset of families across several generations. It contains 101 different families with distinct family names, including 206 nuclear families, 607 individuals, with 14,816 images. The dataset are composed of renowned public families.
Paper: [Kinship Classification by Modeli... | Provide a detailed description of the following dataset: Family101 |
FIW | FIW is a large and comprehensive database available for kinship recognition. FIW is made up of 11,932 natural family photos of 1,000 families-- nearly 10x more than the next-to-largest, [Family-101](family101) database. Also, it contains 656,954 image pairs split between the 11 relationships, which is much larger than ... | Provide a detailed description of the following dataset: FIW |
KinFaceW | KinFaceW consists of two kinship datasets: KinFaceW-I and KinFaceW-II. Face images were collected from the internet, including some public figure face images as well as their parents' or children's face images. Face images are captured under uncontrolled environments in two datasets with no restriction in terms of pose... | Provide a detailed description of the following dataset: KinFaceW |
Boxy | A large vehicle detection dataset with almost two million annotated vehicles for training and evaluating object detection methods for self-driving cars on freeways.
The dataset consists of:
* 200,000 images
* 1,990,000 annotated vehicles
* 5 Megapixel resolution
* Sunshine, rain, dusk, night
* Clear freeways,... | Provide a detailed description of the following dataset: Boxy |
CASR | CASR is a dataset for cyclist arm signal recognition in videos. It contains 219 annotated arm signal actions on videos of approximately 10 seconds each, containing one or two actions per video. | Provide a detailed description of the following dataset: CASR |
Driving Event Camera Dataset | This dataset consists of a number of sequences that were recorded with a VGA (640x480) event camera (Samsung DVS Gen3) and a conventional RGB camera (Huawei P20 Pro) placed on the windshield of a car driving through Zurich. | Provide a detailed description of the following dataset: Driving Event Camera Dataset |
FRIDA | FRIDA and FRIDA2 are databases of numerical images easily usable to evaluate in a systematic way the performance of visibility and contrast restoration algorithms. FRIDA comprises 90 synthetic images of 18 urban road scenes. FRIDA2 comprises 330 synthetic images of 66 diverse road scenes. The view point is closed to th... | Provide a detailed description of the following dataset: FRIDA |
Ford Campus Vision and Lidar Data Set | Ford Campus Vision and Lidar Data Set is a dataset collected by an autonomous ground vehicle testbed, based upon a modified Ford F-250 pickup truck. The vehicle is outfitted with a professional (Applanix POS LV) and consumer (Xsens MTI-G) Inertial Measuring Unit (IMU), a Velodyne 3D-lidar scanner, two push-broom forwar... | Provide a detailed description of the following dataset: Ford Campus Vision and Lidar Data Set |
JAAD | JAAD is a dataset for studying joint attention in the context of autonomous driving. The focus is on pedestrian and driver behaviors at the point of crossing and factors that influence them. To this end, JAAD dataset provides a richly annotated collection of 346 short video clips (5-10 sec long) extracted from over 240... | Provide a detailed description of the following dataset: JAAD |
LISA Vehicle Detection | This is a dataset for vehicle detection. It consists of:
* Three color video sequences captured at different times of the day and illumination settings: morning, evening, sunny, cloudy, etc.
* Different driving environments: highway and urban.
* Varying traffic conditions: light to dense traffic
Paper: [A Gener... | Provide a detailed description of the following dataset: LISA Vehicle Detection |
LLAMAS | The unsupervised Labeled Lane MArkerS dataset (LLAMAS) is a dataset for lane detection and segmentation. It contains over 100,000 annotated images, with annotations of over 100 meters at a resolution of 1276 x 717 pixels. The Unsupervised Llamas dataset was annotated by creating high definition maps for automated drivi... | Provide a detailed description of the following dataset: LLAMAS |
VIsual PERception (VIPER) | VIPER is a benchmark suite for visual perception. The benchmark is based on more than 250K high-resolution video frames, all annotated with ground-truth data for both low-level and high-level vision tasks, including optical flow, semantic instance segmentation, object detection and tracking, object-level 3D scene layou... | Provide a detailed description of the following dataset: VIsual PERception (VIPER) |
REC-COCO | Relations in Captions (REC-COCO) is a new dataset that contains associations between caption tokens and bounding boxes in images. REC-COCO is based on the MS-COCO and V-COCO datasets. For each image in V-COCO, we collect their corresponding captions from MS-COCO and automatically align the concept triplet in V-COCO to ... | Provide a detailed description of the following dataset: REC-COCO |
TRIPOD | TRIPOD contains screenplays and plot synopses with turning point (TP) annotations for 99 movies. Each movie contains:
1. The Wikipedia plot synopsis (extended summary of 35 sentences on average) with sentence-level TP annotations.
2. The screenplay (all dialogue and description parts of the movie) segmented into sc... | Provide a detailed description of the following dataset: TRIPOD |
CSI Screenplay Summarization Corpus | The dataset contains gold-standard summary labels for 39 "CSI: Crime Scene Investigation" episodes from seasons 1-5. Each episode contains the full-length screenplay and human annotations for its summary. The annotations include:
1. scene-level binary labels denoting whether the scene belongs to the summary of the e... | Provide a detailed description of the following dataset: CSI Screenplay Summarization Corpus |
FPV-O | FPV-O is a multi-subject first-person vision dataset of office activities. Office activities include person-to-person interactions, such as chatting and handshaking, person-to-object interactions, such as using a computer or a whiteboard, as well as generic activities such as walking. The videos in the dataset present ... | Provide a detailed description of the following dataset: FPV-O |
MERL Shopping | MERL Shopping is a dataset for training and testing action detection algorithms. The MERL Shopping Dataset consists of 106 videos, each of which is a sequence about 2 minutes long. The videos are from a fixed overhead camera looking down at people shopping in a grocery store setting. Each video contains several instanc... | Provide a detailed description of the following dataset: MERL Shopping |
A2D | A2D (Actor-Action Dataset) is a dataset for simultaneously inferring actors and actions in videos. A2D has seven actor classes (adult, baby, ball, bird, car, cat, and dog) and eight action classes (climb, crawl, eat, fly, jump, roll, run, and walk) not including the no-action class, which we also consider. The A2D has ... | Provide a detailed description of the following dataset: A2D |
ASD | The Annotated Semantic Dataset is composed of $11$ videos, divided in $3$ activity categories: Biking; Driving and Walking, according to their amount of semantic information. The classes are: $0p$, which represents the videos with approximately no semantic information; $25p$, for the videos containing relevant semantic... | Provide a detailed description of the following dataset: ASD |
l2d | This dataset is composed of paired videos of people dancing 3 different music styles: Ballet, Michael Jackson and Salsa.
It contains multimodal data (visual data, temporal-graphs and audio) careful-selected from publicly available videos of dancers performing representative movements of the music style and audio data ... | Provide a detailed description of the following dataset: l2d |
OccludedPASCAL3D+ | The **OccludedPASCAL3D+** is a dataset is designed to evaluate the robustness to occlusion for a number of computer vision tasks, such as object detection, keypoint detection and pose estimation. In the OccludedPASCAL3D+ dataset, we simulate partial occlusion by superimposing objects cropped from the MS-COCO dataset on... | Provide a detailed description of the following dataset: OccludedPASCAL3D+ |
THEODORE | Recent work about synthetic indoor datasets from perspective views has shown significant improvements of object detection results with Convolutional Neural Networks(CNNs). In this paper, we introduce THEODORE: a novel, large-scale indoor dataset containing 100,000 high- resolution diversified fisheye images with 14 cla... | Provide a detailed description of the following dataset: THEODORE |
MHRI dataset | The dataset includes recordings from 10 different users teaching the robot different common kitchen objects, that consists of synchronized recordings from three cameras and a microphone mounted on the robot:
An RGB-d camera covers the user manipulation and interaction with the robot
An RGB-d camera mounted ... | Provide a detailed description of the following dataset: MHRI dataset |
highD Dataseth | The highD dataset is a new dataset of naturalistic vehicle trajectories recorded on German highways. Using a drone, typical limitations of established traffic data collection methods such as occlusions are overcome by the aerial perspective. Traffic was recorded at six different locations and includes more than 110 500... | Provide a detailed description of the following dataset: highD Dataseth |
inD Dataset | The **inD** dataset is a new dataset of naturalistic vehicle trajectories recorded at German intersections. Using a drone, typical limitations of established traffic data collection methods like occlusions are overcome. Traffic was recorded at four different locations. The trajectory for each road user and its type is ... | Provide a detailed description of the following dataset: inD Dataset |
rounD Dataset | The rounD dataset is a new dataset of naturalistic road user trajectories recorded at German roundabouts. Using a drone, typical limitations of established traffic data collection methods like occlusions are overcome. Traffic was recorded at three different locations. The trajectory for each road user and its type is e... | Provide a detailed description of the following dataset: rounD Dataset |
Localized Narratives | We propose Localized Narratives, a new form of multimodal image annotations connecting vision and language. We ask annotators to describe an image with their voice while simultaneously hovering their mouse over the region they are describing. Since the voice and the mouse pointer are synchronized, we can localize every... | Provide a detailed description of the following dataset: Localized Narratives |
CE4 | Given the difficulty to handle planetary data we provide downloadable files in PNG format from the missions Chang'E-3 and Chang'E-4. In addition to a set of scripts to do the conversion given a different PDS4 Dataset.
This set of images constitute one of the first available datasets to tackle problems of Computer V... | Provide a detailed description of the following dataset: CE4 |
MICC-SRI | The dataset contains 11,913 frame pairs of urban driving footage with and without moving objects, synthetically generated with the CARLA simulator. All frames are available both as RGB images and semantic segmentations. RGB images are non-photorealistic being rendered by a game engine, while semantic segmentations are ... | Provide a detailed description of the following dataset: MICC-SRI |
KITTI-trajectory-prediction | KITTI is a well established dataset in the computer vision community. It has often been used for trajectory prediction despite not having a well defined split, generating non comparable baselines in different works. This dataset aims at bridging this gap and proposes a well defined split of the KITTI data.
Samples are... | Provide a detailed description of the following dataset: KITTI-trajectory-prediction |
EmoContext | EmoContext consists of three-turn English Tweets. The emotion labels include happiness, sadness, anger and other. | Provide a detailed description of the following dataset: EmoContext |
Glint360K | The largest and cleanest face recognition dataset Glint360K,
which contains **`17,091,657`** images of **`360,232`** individuals, baseline models trained on Glint360K can easily achieve state-of-the-art performance. | Provide a detailed description of the following dataset: Glint360K |
IndicCorp | IndicCorp is a large monolingual corpora with around 9 billion tokens covering 12 of the major Indian languages. It has been developed by discovering and scraping thousands of web sources - primarily news, magazines and books, over a duration of several months.
**Languages covered**: Assamese, Bengali, English, Guja... | Provide a detailed description of the following dataset: IndicCorp |
RuFa | RuFa (Ruqaa-Farsi) dataset contains images of text written in one of two Arabic fonts: Ruqaa and Nastaliq (Farsi). The dataset contains 40,000 synthesized image and 516 real one, 40,516 in total. Images are in RGB JPG format at 100×100px. Text in the images has varying number of words, position, size, and opacity.
R... | Provide a detailed description of the following dataset: RuFa |
MERL-RAV | The MERL-RAV (MERL Reannotation of AFLW with Visibility) Dataset contains over 19,000 face images in a full range of head poses. Each face is manually labeled with the ground-truth locations of 68 landmarks, with the additional information of whether each landmark is unoccluded, self-occluded (due to extreme head poses... | Provide a detailed description of the following dataset: MERL-RAV |
News Interactions on Globo.com | ### Context
This large dataset with users interactions logs (page views) from a news portal was kindly provided by [Globo.com][1], the most popular news portal in Brazil, for reproducibility of the experiments with CHAMELEON - a meta-architecture for contextual hybrid session-based news recommender systems. The sour... | Provide a detailed description of the following dataset: News Interactions on Globo.com |
Synbols | Synbols is a dataset generator designed for probing the behavior of learning algorithms. By defining the distribution over latent factors one can craft a dataset specifically tailored to answer specific questions about a given algorithm.
Default versions of these datasets are also materialized and can serve as bench... | Provide a detailed description of the following dataset: Synbols |
C&Z | One of the first datasets (if not the first) to highlight the importance of bias and diversity in the community, which started a revolution afterwards. Introduced in 2014 as integral part of a thesis of Master of Science [1,2] at Carnegie Mellon and City University of Hong Kong. It was later expanded by adding syntheti... | Provide a detailed description of the following dataset: C&Z |
GEM | Generation, Evaluation, and Metrics (GEM) is a benchmark environment for Natural Language Generation with a focus on its Evaluation, both through human annotations and automated Metrics.
GEM aims to:
- measure NLG progress across 13 datasets spanning many NLG tasks and languages.
- provide an in-depth analysis o... | Provide a detailed description of the following dataset: GEM |
ALFWorld | ALFWorld contains interactive TextWorld environments (Côté et. al) that parallel embodied worlds in the ALFRED dataset (Shridhar et. al). The aligned environments allow agents to reason and learn high-level policies in an abstract space before solving embodied tasks through low-level actuation. | Provide a detailed description of the following dataset: ALFWorld |
HQ-WMCA | The High-Quality Wide Multi-Channel Attack database (HQ-WMCA) database consists of 2904 short multi-modal video recordings of both bona-fide and presentation attacks. There are 555 bonafide presentations from 51 participants and the remaining 2349 are presentation attacks. The data is recorded from several channels inc... | Provide a detailed description of the following dataset: HQ-WMCA |
The Best Sarcasm Annotated Dataset in Spanish | ### Content
This dataset contains all utterances of two episodes of South Park (Latin American voices) and two episodes of Archer (Spanish voices). The order of the utterances is shuffled. Each utterance has been annotated based on whether it is sarcastic or not. Sarcastic expressions also contain further annotation... | Provide a detailed description of the following dataset: The Best Sarcasm Annotated Dataset in Spanish |
MIRACL-VC1 | MIRACL-VC1 is a lip-reading dataset including both depth and color images. It can be used for diverse research fields like visual speech recognition, face detection, and biometrics. Fifteen speakers (five men and ten women) positioned in the frustum of an MS Kinect sensor and utter ten times a set of ten words and ten ... | Provide a detailed description of the following dataset: MIRACL-VC1 |
XD-Violence | XD-Violence is a large-scale audio-visual dataset for violence detection in videos. | Provide a detailed description of the following dataset: XD-Violence |
PatentMatch | We address the computer-assisted search for prior art by creating a training dataset for supervised machine learning called PatentMatch. It contains pairs of claims from patent applications and semantically corresponding text passages of different degrees from cited patent documents. Each pair has been labeled by techn... | Provide a detailed description of the following dataset: PatentMatch |
A Dataset of Journalists' Interactions with Their Readership | We present a dataset of dialogs in which journalists of The Guardian replied to reader comments and identify the reasons why. Based on this data, we formulate the novel task of recommending reader comments to journalists that are worth reading or replying to, i.e., ranking comments in such a way that the top comments a... | Provide a detailed description of the following dataset: A Dataset of Journalists' Interactions with Their Readership |
Top Comment or Flop Comment? | This dataset comprises four files of IDs of either strongly or weakly engaging online news comments (please see the paper for details):
"Top comments" are 1) the top 10% comments in the politics section of The Guardian with the largest relative number of *replies* received (3111 samples) and 2) the top 10% comments in... | Provide a detailed description of the following dataset: Top Comment or Flop Comment? |
HeartSeg | The medaka (Oryzias latipes) and the zebrafish (Danio rerio) are used as a model organism for a variety of subjects in biomedical research. The presented work aims to study the potential of automated ventricular dimension estimation through heart segmentation in medaka. For more on this, it's time for a closer look on ... | Provide a detailed description of the following dataset: HeartSeg |
DNS Challenge | The DNS Challenge at INTERSPEECH 2020 intended to promote collaborative research in single-channel Speech Enhancement aimed to maximize the perceptual quality and intelligibility of the enhanced speech. The challenge evaluated the speech quality using the online subjective evaluation framework ITU-T P.808. The challeng... | Provide a detailed description of the following dataset: DNS Challenge |
Interspeech 2021 Deep Noise Suppression Challenge | The Deep Noise Suppression (DNS) challenge is designed to foster innovation in the area of noise suppression to achieve superior perceptual speech quality.
This challenge has two two tracks:
**Track 1: Real-Time Denoising track for wide band scenario**
The noise suppressor must take less than the stride time T... | Provide a detailed description of the following dataset: Interspeech 2021 Deep Noise Suppression Challenge |
TRN | The Toulouse Road Network dataset describes patches of road maps from the city of Toulouse, represented both as spatial graphs G = (A, X) and as grayscale segmentation images.
The TRN dataset contains 111,034 data points (map tiles), of which: 80,357 are in the training set (around 72.4%), 11,679 are in the validat... | Provide a detailed description of the following dataset: TRN |
WEB-FORUM-52 | The WEB-FORUM-52 gold standard comprises (i) 13 web forums from the health domain, (ii) 15 forums obtained from a Wikipedia list of popular forums (https://en.wikipedia.org/wiki/List_of_Internet_forums), (iii) 13 forums mentioned on a list of popular German Web forums (https://www.beliebte-foren.de), (iv) nine forums o... | Provide a detailed description of the following dataset: WEB-FORUM-52 |
KorQuAD | KorQuAD is a large-scale question-and-answer dataset constructed for Korean machine reading comprehension, and investigate the dataset to understand the distribution of answers and the types of reasoning required to answer the question. This dataset benchmarks the data generating process of SQuAD to meet the standard. | Provide a detailed description of the following dataset: KorQuAD |
MOBIO | The MOBIO database consists of bi-modal (audio and video) data taken from 152 people. The database has a female-male ratio or nearly 1:2 (100 males and 52 females) and was collected from August 2008 until July 2010 in six different sites from five different countries. This led to a diverse bi-modal database with both n... | Provide a detailed description of the following dataset: MOBIO |
FRLL-Morphs | FRLL-Morphs is a dataset of morphed faces based on images selected from the publicly available Face Research London Lab dataset [1].
We created the database by selecting similar looking pairs of people, and made 4 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGA... | Provide a detailed description of the following dataset: FRLL-Morphs |
VisualMRC | VisualMRC is a visual machine reading comprehension dataset that proposes a task: given a question and a document image, a model produces an abstractive answer.
You can find more details, analyses, and baseline results in the paper,
VisualMRC: Machine Reading Comprehension on Document Images, AAAI 2021.
Stati... | Provide a detailed description of the following dataset: VisualMRC |
FERET-Morphs | FERET-Morphs is a dataset of morphed faces selected from the publicly available FERET dataset [1].
We created the database by selecting similar looking pairs of people, and made 3 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGAN 2 [3].
* [1] https://www.nist.... | Provide a detailed description of the following dataset: FERET-Morphs |
FRGC-Morphs | FRGC-Morphs is a dataset of morphed faces selected from the publicly available FRGC dataset [1].
We created the database by selecting similar looking pairs of people, and made 3 types of morphs for each pair using the following morphing tools: OpenCV [2], FaceMorpher [3], StyleGAN 2 [3].
* [1] https://www.nist.go... | Provide a detailed description of the following dataset: FRGC-Morphs |
NISP- A Multi-lingual Multi-accent Dataset for Speaker Profiling | We announce the release of a new multilingual speaker dataset called NITK-IISc Multilingual Multi-accent Speaker Profiling(NISP) dataset. The dataset contains speech in six different languages -- five Indian languages along with Indian English. The dataset contains speech data from 345 bilingual speakers in India. Each... | Provide a detailed description of the following dataset: NISP- A Multi-lingual Multi-accent Dataset for Speaker Profiling |
NinaPro DB2 | The second Ninapro database includes 40 intact subjects and it is thoroughly described in the paper: "Manfredo Atzori, Arjan Gijsberts, Claudio Castellini, Barbara Caputo, Anne-Gabrielle Mittaz Hager, Simone Elsig, Giorgio Giatsidis, Franco Bassetto & Henning Müller. Electromyography data for non-invasive naturally-con... | Provide a detailed description of the following dataset: NinaPro DB2 |
BuzzFeed-Webis Fake News Corpus 2016 | The BuzzFeed-Webis Fake News Corpus 16 comprises the output of 9 publishers in a week close to the US elections. Among the selected publishers are 6 prolific hyperpartisan ones (three left-wing and three right-wing), and three mainstream publishers (see Table 1). All publishers earned Facebook’s blue checkmark, indicat... | Provide a detailed description of the following dataset: BuzzFeed-Webis Fake News Corpus 2016 |
POLIT-FALSE-n-LEGIT NEWS DB 2016-2017 | The LiT.RL POLIT-FALSE-n-LEGIT NEWS DB 2016-2017 contains a total of 274 news articles about U.S. Politics, content-matched in pairs of legitimate and falsified news. The database is free and released under an open license for educational and research purposes. | Provide a detailed description of the following dataset: POLIT-FALSE-n-LEGIT NEWS DB 2016-2017 |
GQN rooms-ring-camera | GQN rooms-ring-camera consist of scenes of a variable number of random objects captured in a square room of size 7x7 units. Wall textures, floor textures as well as the shapes of the objects are randomly chosen within a fixed pool of discrete options. There are 5 possible wall textures (red, green, cerise, orange, yell... | Provide a detailed description of the following dataset: GQN rooms-ring-camera |
ISOT Fake News Dataset | The ISOT Fake News dataset is a compilation of several thousands fake news and truthful articles, obtained from different legitimate news sites and sites flagged as unreliable by Politifact.com. | Provide a detailed description of the following dataset: ISOT Fake News Dataset |
ObjectsRoom | The **ObjectsRoom** dataset is based on the MuJoCo environment used by the Generative Query Network [4] and is a multi-object extension of the 3d-shapes dataset. The training set contains 1M scenes with up to three objects. We also provide ~1K test examples for the following variants:
2.1 Empty room: scenes consist ... | Provide a detailed description of the following dataset: ObjectsRoom |
SVDC Fake News Dataset | A labeled dataset that presents fake news surrounding the conflict in Syria. The dataset consists of a set of articles/news labeled by 0 (fake) or 1 (credible). Credibility of articles are computed with respect to a ground truth information obtained from the Syrian Violations Documentation Center (VDC). In particular,... | Provide a detailed description of the following dataset: SVDC Fake News Dataset |
FakeNewsAMT & Celebrity | **FakeNewsAMT & Celebrity** include two novel datasets for the task of fake news detection, covering seven different news domains. | Provide a detailed description of the following dataset: FakeNewsAMT & Celebrity |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.