dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
AV Digits Database
AV Digits Database is an audiovisual database which contains normal, whispered and silent speech. 53 participants were recorded from 3 different views (frontal, 45 and profile) pronouncing digits and phrases in three speech modes. The database consists of two parts: digits and short phrases. In the first part, parti...
Provide a detailed description of the following dataset: AV Digits Database
Fabrics Dataset
The Fabrics Dataset consists of about 2000 samples of garments and fabrics. A small patch of each surface has been captured under 4 different illumination conditions using a custom made, portable photometric stereo sensor. All images have been acquired "in the field" (at clothes shops) and the dataset reflects the dist...
Provide a detailed description of the following dataset: Fabrics Dataset
MobiFace
MobiFace is the first dataset for single face tracking in mobile situations. It consists of 80 unedited live-streaming mobile videos captured by 70 different smartphone users in fully unconstrained environments. Over 95K bounding boxes are manually labelled. The videos are carefully selected to cover typical smartphone...
Provide a detailed description of the following dataset: MobiFace
LSFM
The Large Scale Facial Model (LSFM) is a 3D statistical model of facial shape built from nearly 10,000 individuals.
Provide a detailed description of the following dataset: LSFM
FaceScape
FaceScape dataset provides 3D face models, parametric models and multi-view images in large-scale and high-quality. The camera parameters, the age and gender of the subjects are also included. The data have been released to public for non-commercial research purpose.
Provide a detailed description of the following dataset: FaceScape
AgeDB
AgeDB contains 16, 488 images of various famous people, such as actors/actresses, writers, scientists, politicians, etc. Every image is annotated with respect to the identity, age and gender attribute. There exist a total of 568 distinct subjects. The average number of images per subject is 29. The minimum and maximum ...
Provide a detailed description of the following dataset: AgeDB
AFEW-VA
The AFEW-VA databaset is a collection of highly accurate per-frame annotations levels of valence and arousal, along with per-frame annotations of 68 facial landmarks for 600 challenging video clips. These clips are extracted from feature films and were also annotated in terms of discrete emotion categories in the form ...
Provide a detailed description of the following dataset: AFEW-VA
KILT
**KILT** (**Knowledge Intensive Language Tasks**) is a benchmark consisting of 11 datasets representing 5 types of tasks: * Fact-checking (FEVER), * Entity linking (AIDA CoNLL-YAGO, WNED-WIKI, WNED-CWEB), * Slot filling (T-Rex, Zero Shot RE), * Open domain QA (Natural Questions, HotpotQA, TriviaQA, ELI5), * Dial...
Provide a detailed description of the following dataset: KILT
SOREL-20M
SOREL-20M is a large-scale dataset consisting of nearly 20 million files with pre-extracted features and metadata, high-quality labels derived from multiple sources, information about vendor detections of the malware samples at the time of collection, and additional “tags” related to each malware sample to serve as add...
Provide a detailed description of the following dataset: SOREL-20M
Relational Pattern Similarity Dataset
The relational pattern similarity dataset is a new dataset upon the work of Zeichner et al. (2012), which consists of relational patterns with semantic inference labels annotated. The dataset includes 5,555 pairs extracted by Reverb (Fader et al., 2011), 2,447 pairs with inference relation and 3,108 pairs (the rest) wi...
Provide a detailed description of the following dataset: Relational Pattern Similarity Dataset
PHM2017
PHM2017 is a new dataset consisting of 7,192 English tweets across six diseases and conditions: Alzheimer’s Disease, heart attack (any severity), Parkinson’s disease, cancer (any type), Depression (any severity), and Stroke. The Twitter search API was used to retrieve the data using the colloquial disease names as sear...
Provide a detailed description of the following dataset: PHM2017
ORVS
The ORVS dataset has been newly established as a collaboration between the computer science and visual-science departments at the University of Calgary. This dataset contains 49 images (42 training and seven testing images) collected from a clinic in Calgary-Canada. All images were acquired with a Zeiss Visucam 200 ...
Provide a detailed description of the following dataset: ORVS
DR HAGIS
The DR HAGIS database has been created to aid the development of vessel extraction algorithms suitable for retinal screening programmes. Researchers are encouraged to test their segmentation algorithms using this database. All thirty-nine fundus images were obtained from a diabetic retinopathy screening programme in...
Provide a detailed description of the following dataset: DR HAGIS
ARIA
This data set was collected in 2004 to 2006 in the United Kingdom. Subjects were adult males and females, some of whom were healthy (control group), some with age-related macular degeneration (AMD group), and some were diabetic patients (diabetic group). Unfortunately, no other information from this time exists about t...
Provide a detailed description of the following dataset: ARIA
VICAVR
The VICAVR database is a set of retinal images used for the computation of the A/V Ratio. The database currently includes 58 images. The images have been acquired with a TopCon non-mydriatic camera NW-100 model and are optic disc centered with a resolution of 768x584. The database includes the caliber of the vessels me...
Provide a detailed description of the following dataset: VICAVR
OCTAGON
The OCTAGON dataset is a set of Angiography by Octical Coherence Tomography images (OCT-A) used to the segmentation of the Foveal Avascular Zone (FAZ). The dataset includes 144 healthy OCT-A images and 69 diabetic OCT-A images, divided into four groups, each one with 36 and about 17 OCT-A images, respectively. These gr...
Provide a detailed description of the following dataset: OCTAGON
CLOUD
The CLOUD dataset is a set of Optical Coherence Tomography of the Anterior Segment images (AS-OCT) used to the automatic identification and representation of the cornea-contact lens relationship. The dataset includes 112 AS-OCT images that were captured from 16 different patients. In particular, the images were obtaine...
Provide a detailed description of the following dataset: CLOUD
MESSIDOR
The Messidor database has been established to facilitate studies on computer-assisted diagnoses of diabetic retinopathy. The research community is welcome to test its algorithms on this database. In this section, you will find instructions on how to download the database.
Provide a detailed description of the following dataset: MESSIDOR
DIARETDB1
The database consists of 89 colour fundus images of which 84 contain at least mild non-proliferative signs (Microaneurysms) of the diabetic retinopathy, and 5 are considered as normal which do not contain any signs of the diabetic retinopathy according to all experts who participated in the evaluation. Images were capt...
Provide a detailed description of the following dataset: DIARETDB1
UDA-CH
UDA-CH contains 16 objects that cover a variety of artworks which can be found in a museum like sculptures, paintings and books. Specifically, the dataset has been collected inside the cultural site “Galleria Regionale di Palazzo Bellomo” located in Siracusa, Italy.
Provide a detailed description of the following dataset: UDA-CH
EGO-CH
EGO-CH is a dataset of egocentric videos for visitors’ behavior understanding. The dataset has been collected in two different cultural sites and includes more than 27 hours of video acquired by 70 subjects, including volunteers and 60 real visitors. The overall dataset includes labels for 26 environments and over 200 ...
Provide a detailed description of the following dataset: EGO-CH
MAP
**Maybe Ambiguous Pronoun** is a dataset similar to [GAP](/dataset/gap-coreference-dataset) dataset, but without binary gender constraints.
Provide a detailed description of the following dataset: MAP
GICoref
GICoref is a fully annotated coreference resolution dataset written by and about trans people.
Provide a detailed description of the following dataset: GICoref
NAF
This dataset was created with images provided by the United States National Archive and FamilySearch. The goal of this data is to capture relationships between text/handwriting entities on form images. It will include transcriptions in the future, but doesn't currently. The form images are organized into "groups"...
Provide a detailed description of the following dataset: NAF
ImageNet-P
**ImageNet-P** consists of noise, blur, weather, and digital distortions. The dataset has validation perturbations; has difficulty levels; has CIFAR-10, Tiny ImageNet, ImageNet 64 × 64, standard, and Inception-sized editions; and has been designed for benchmarking not training networks. ImageNet-P departs from ImageNet...
Provide a detailed description of the following dataset: ImageNet-P
Combinatorial 3D Shape Dataset
The combinatorial 3D shape dataset is composed of 406 instances of 14 classes. Specifically, each object in the dataset is considered equivalent to a sequence of primitive placement.
Provide a detailed description of the following dataset: Combinatorial 3D Shape Dataset
AI2D
AI2 Diagrams (AI2D) is a dataset of over 5000 grade school science diagrams with over 150000 rich annotations, their ground truth syntactic parses, and more than 15000 corresponding multiple choice questions.
Provide a detailed description of the following dataset: AI2D
Chart2Text
Chart2Text is a dataset that was crawled from 23,382 freely accessible pages from statista.com in early March of 2020, yielding a total of 8,305 charts, and associated summaries. For each chart, the chart image, the underlying data table, the title, the axis labels, and a human-written summary describing the statistic ...
Provide a detailed description of the following dataset: Chart2Text
DENSE
DENSE (Depth Estimation oN Synthetic Events) is a new dataset with synthetic events and perfect ground truth.
Provide a detailed description of the following dataset: DENSE
PixelShift200
Advanced pixel shift technology is employed to perform a full color sampling of the image. Pixel shift technology takes four samples of the same image at nearly the same time, and physically controls the camera sensor to move one pixel horizontally or vertically at each sampling to capture all color information at each...
Provide a detailed description of the following dataset: PixelShift200
VLEP
VLEP contains 28,726 future event prediction examples (along with their rationales) from 10,234 diverse TV Show and YouTube Lifestyle Vlog video clips. Each example (see Figure 1) consists of a Premise Event (a short video clip with dialogue), a Premise Summary (a text summary of the premise event), and two potential n...
Provide a detailed description of the following dataset: VLEP
Cata7
Cata7 is the first cataract surgical instrument dataset for semantic segmentation. The dataset consists of seven videos while each video records a complete cataract surgery. All videos are from Beijing Tongren Hospital. Each video is split into a sequence of images, where resolution is 1920×1080 pixels. To reduce redun...
Provide a detailed description of the following dataset: Cata7
UCC
The Unhealthy Comments Corpus (UCC) is corpus of 44355 comments intended to assist in research on identifying subtle attributes which contribute to unhealthy conversations online. Each comment is labelled as either 'healthy' or 'unhealthy', in addition to binary labels for the presence of six potentially 'unhealthy'...
Provide a detailed description of the following dataset: UCC
Satire Dataset
The satire dataset is a new multi-modal dataset of satirical and regular news articles. The satirical news is collected from four websites that explicitly declare themselves to be satire, and the regular news is collected from six mainstream news websites. Specifically, the satirical news websites the articles were col...
Provide a detailed description of the following dataset: Satire Dataset
Headcam
This dataset contains panoramic video captured from a helmet-mounted camera while riding a bike through suburban Northern Virginia.
Provide a detailed description of the following dataset: Headcam
OCNLI
OCNLI stands for Original Chinese Natural Language Inference. It is corpus for Chinese Natural Language Inference, collected following closely the procedures of MNLI, but with enhanced strategies aiming for more challenging inference pairs. No human/machine translation is used in creating the dataset, and thus the Chin...
Provide a detailed description of the following dataset: OCNLI
QReCC
QReCC contains 14K conversations with 81K question-answer pairs. QReCC is built on questions from TREC CAsT, QuAC and Google Natural Questions. While TREC CAsT and QuAC datasets contain multi-turn conversations, Natural Questions is not a conversational dataset. Questions in NQ dataset were used as prompts to create co...
Provide a detailed description of the following dataset: QReCC
PHD²
The dataset contains information on what video segments a specific user considers a highlight. Having this kind of data allows for strong personalization models, as specific examples of what a user is interested in help models obtain a fine-grained understanding of that specific user. The data consists of YouTube vi...
Provide a detailed description of the following dataset: PHD²
Video2GIF
The **Video2GIF** dataset contains over 100,000 pairs of GIFs and their source videos. The GIFs were collected from two popular GIF websites (makeagif.com, gifsoup.com) and the corresponding source videos were collected from YouTube in Summer 2015. IDs and URLs of the GIFs and the videos are provided, along with tempor...
Provide a detailed description of the following dataset: Video2GIF
VAST
VAST consists of a large range of topics covering broad themes, such as politics (e.g., ‘a Palestinian state’), education (e.g., ‘charter schools’), and public health (e.g., ‘childhood vaccination’). In addition, the data includes a wide range of similar expressions (e.g., ‘guns on campus’ versus ‘firearms on campus’)....
Provide a detailed description of the following dataset: VAST
Silent Speech EMG
Facial electromyography recordings during both silent and vocalized speech.
Provide a detailed description of the following dataset: Silent Speech EMG
SMOT
The SMOT dataset, Single sequence-Multi Objects Training, is collected to represent a practical scenario of collecting training images of new objects in the real world, i.e. a mobile robot with an RGB-D camera collects a sequence of frames while driving around a table to learning multiple objects and tries to recognize...
Provide a detailed description of the following dataset: SMOT
3DNet
The 3DNet dataset is a free resource for object class recognition and 6DOF pose estimation from point cloud data. 3DNet provides a large-scale hierarchical CAD-model databases with increasing numbers of classes and difficulty with 10, 60 and 200 object classes together with evaluation datasets that contain thousands of...
Provide a detailed description of the following dataset: 3DNet
ARID
ARID is a large-scale, multi-view object dataset collected with an RGB-D camera mounted on a mobile robot.
Provide a detailed description of the following dataset: ARID
OCID
Developing robot perception systems for handling objects in the real-world requires computer vision algorithms to be carefully scrutinized with respect to the expected operating domain. This demands large quantities of ground truth data to rigorously evaluate the performance of algorithms. The Object Cluttered Indoo...
Provide a detailed description of the following dataset: OCID
LfED-6D
The LfED-6D dataset is a collection of 6D grasp annotations acquired through experience (with a robot platform) or by human demonstration. For known objects, the annotated grasps can be directly applied given the pose of the object model is correctly computed. For unknown objects, the grasps can be generalized using me...
Provide a detailed description of the following dataset: LfED-6D
NYU-VP
NYU-VP is a new dataset for multi-model fitting, vanishing point (VP) estimation in this case. Each image is annotated with up to eight vanishing points, and pre-extracted line segments are provided which act as data points for a robust estimator. Due to its size, the dataset is the first to allow for supervised learni...
Provide a detailed description of the following dataset: NYU-VP
YUD+
YUD+ is a dataset containing additional Vanishing Point Labels for the [York Urban Database](https://paperswithcode.com/dataset/york-urban-line-segment-database).
Provide a detailed description of the following dataset: YUD+
NText
NText is an eight million words dataset extracted and preprocessed from nuclear research papers and thesis.
Provide a detailed description of the following dataset: NText
NQuAD
NQuAD is a Nuclear Question Answering Dataset, which contains 700+ nuclear Question Answer pairs developed and verified by expert nuclear researchers.
Provide a detailed description of the following dataset: NQuAD
Indoor and outdoor DFD dataset
The dfd_indoor dataset contains 110 images for training and 29 images for testing. The dfd_outdoor dataset contains 34 images for tests; no ground truth was given for this dataset, as the depth sensor only works on indoor scenes.
Provide a detailed description of the following dataset: Indoor and outdoor DFD dataset
Lorenz Dataset
The Lorenz dataset contains 100000 time-series with length 24. The data has 5 modes and it is obtained using the Lorenz equation with 5 different seed values.
Provide a detailed description of the following dataset: Lorenz Dataset
EHR-Rel
EHR-RelB is a benchmark dataset for biomedical concept relatedness, consisting of 3630 concept pairs sampled from electronic health records (EHRs). EHR-RelA is a smaller dataset of 111 concept pairs, which are mainly unrelated.
Provide a detailed description of the following dataset: EHR-Rel
MLGESTURE DATASET
MlGesture is a dataset for hand gesture recognition tasks, recorded in a car with 5 different sensor types at two different viewpoints. The dataset contains over 1300 hand gesture videos from 24 participants and features 9 different hand gesture symbols. One sensor cluster with five different cameras is mounted in fron...
Provide a detailed description of the following dataset: MLGESTURE DATASET
NYT-H
NYT-H is a dataset for distantly-supervised relation extraction, in which DS-labelled training data is used and several annotators to label test data are hired. NYT-H can serve as a benchmark of distantly-supervised relation extraction.
Provide a detailed description of the following dataset: NYT-H
CSAW-S
CSAW-S is a dataset of mammography images which includes expert annotations of tumors and non-expert annotations of breast anatomy and artifacts in the image.
Provide a detailed description of the following dataset: CSAW-S
2D-3D Match Dataset
2D-3D Match Dataset is a new dataset of 2D-3D correspondences by leveraging the availability of several 3D datasets from RGB-D scans. Specifically, the data from SceneNN and 3DMatch are used. The training dataset consists of 110 RGB-D scans, of which 56 scenes are from SceneNN and 54 scenes are from 3DMatch. The 2D-3D ...
Provide a detailed description of the following dataset: 2D-3D Match Dataset
FIGR-8
The FIGR-8 database is a dataset containing 17,375 classes of 1,548,256 images representing pictograms, ideograms, icons, emoticons or object or conception depictions. Its aim is to set a benchmark for Few-shot Image Generation tasks, albeit not being limited to it. Each image is represented by 192x192 pixels with gray...
Provide a detailed description of the following dataset: FIGR-8
ImageNet-Sketch
ImageNet-Sketch data set consists of 50,889 images, approximately 50 images for each of the 1000 ImageNet classes. The data set is constructed with Google Image queries "sketch of __", where __ is the standard class name. Only within the "black and white" color scheme is searched. 100 images are initially queried for ...
Provide a detailed description of the following dataset: ImageNet-Sketch
YouTube-VIS 2019
YouTubeVIS is a new dataset tailored for tasks like simultaneous detection, segmentation and tracking of object instances in videos and is collected based on the current largest video object segmentation dataset YouTubeVOS.
Provide a detailed description of the following dataset: YouTube-VIS 2019
EURLEX57K
EURLEX57K is a new publicly available legal LMTC dataset, dubbed EURLEX57K, containing 57k English EU legislative documents from the EUR-LEX portal, tagged with ∼4.3k labels (concepts) from the European Vocabulary (EUROVOC).
Provide a detailed description of the following dataset: EURLEX57K
Anonymized Keystrokes Dataset
Includes two datasets for this task, one for English-French (En-Fr) and another for English-German (En-De). For each dataset, the action sequences for full documents are provided, along with an editor identifier. The dataset contains document-level post-editing action sequences, including edit operations from keystroke...
Provide a detailed description of the following dataset: Anonymized Keystrokes Dataset
METU-VIREF Dataset
**METU-VIREF** is a video referring expression dataset comprising of videos from VIRAT Ground and ILSVRC2015 VID datasets. VIRAT is a surveillance dataset and contains mainly people and vehicles. To line up with this and restrict the domain, only videos that contain vehicles from the ILSVRC dataset are used. The METU-V...
Provide a detailed description of the following dataset: METU-VIREF Dataset
UNDD
UNDD consists of 7125 unlabelled day and night images; additionally, it has 75 night images with pixel-level annotations having classes equivalent to Cityscapes dataset.
Provide a detailed description of the following dataset: UNDD
Mapillary Vistas Dataset
Mapillary Vistas Dataset is a diverse street-level imagery dataset with pixel‑accurate and instance‑specific human annotations for understanding street scenes around the world.
Provide a detailed description of the following dataset: Mapillary Vistas Dataset
Food.com Recipes and Interactions
Food.com Recipes and Interactions consists of 270K recipes and 1.4M user-recipe interactions (reviews) scraped from Food.com, covering a period of 18 years (January 2000 to December 2018).
Provide a detailed description of the following dataset: Food.com Recipes and Interactions
SYNTHIA-PANO
SYNTHIA-PANO is the panoramic version of SYNTHIA dataset. Five sequences are included: Seqs02-summer, Seqs02-fall, Seqs04-summer, Seqs04-fall and Seqs05-summer. Panomaramic images with fine annotation for semantic segmentation.
Provide a detailed description of the following dataset: SYNTHIA-PANO
SUIM
The Segmentation of Underwater IMagery (SUIM) dataset contains over 1500 images with pixel annotations for eight object categories: fish (vertebrates), reefs (invertebrates), aquatic plants, wrecks/ruins, human divers, robots, and sea-floor. The images have been rigorously collected during oceanic explorations and huma...
Provide a detailed description of the following dataset: SUIM
CSAbstruct Dataset
CSAbstruct is a new dataset of annotated computer science abstracts with sentence labels according to their rhetorical roles. The key difference between this dataset and PUBMED-RCT is that PubMed abstracts are written according to a predefined structure, whereas computer science papers are free-form. Therefore, there i...
Provide a detailed description of the following dataset: CSAbstruct Dataset
Pesteh-Set
Pesteh-Set is made of two parts. The first part includes 423 images with ground truth. The pistachios are sorted into two classes: Open-mouth and closed-mouth. The ground truth of the images is a CSV file that consists of the bounding boxes of the two classes of pistachios in the images. There are between 1 to 27 pista...
Provide a detailed description of the following dataset: Pesteh-Set
CelebAGaze
CelebAGaze consists of 25283 high-resolution celebrity images that are collected from CelebA and the Internet. It consists of 21832 face images with eyes staring at the camera and 3451 face images with eyes staring somewhere else. All images (256 × 256) are cropped and the eye mask region by dlib is computed. Specifica...
Provide a detailed description of the following dataset: CelebAGaze
IQUAD
IQUAD is a dataset for Visual Question Answering in interactive environments. It is built upon AI2-THOR, a simulated photo-realistic environment of configurable indoor scenes with interactive object. IQUAD V1 has 75,000 questions, each paired with a unique scene configuration.
Provide a detailed description of the following dataset: IQUAD
EVE
EVE (End-to-end Video-based Eye-tracking) is a dataset for eye-tracking. It is collected from 54 participants and consists of 4 camera views, over 12 million frames and 1327 unique visual stimuli (images, video, text), adding up to approximately 105 hours of video data in total. Official competition on Codalab: [htt...
Provide a detailed description of the following dataset: EVE
OpoSum
OPOSUM is a dataset for the training and evaluation of Opinion Summarization models which contains Amazon reviews from six product domains: Laptop Bags, Bluetooth Headsets, Boots, Keyboards, Televisions, and Vacuums. The six training collections were created by downsampling from the Amazon Product Dataset introduced i...
Provide a detailed description of the following dataset: OpoSum
ForecastQA
ForecastQA is a question-answering dataset consisting of 10,392 event forecasting questions, which have been collected and verified via crowdsourcing efforts. The forecasting problem for this dataset is formulated as a restricted-domain, multiple-choice, question-answering (QA) task that simulates the forecasting scena...
Provide a detailed description of the following dataset: ForecastQA
TSU
Toyota Smarthome Untrimmed (TSU) is a dataset for activity detection in long untrimmed videos. The dataset contains 536 videos with an average duration of 21 mins. Since this dataset is based on the same footage video as Toyota Smarthome Trimmed version, it features the same challenges and introduces additional ones. T...
Provide a detailed description of the following dataset: TSU
AUTSL
The Ankara University Turkish Sign Language Dataset (AUTSL) is a large-scale, multimode dataset that contains isolated Turkish sign videos. It contains 226 signs that are performed by 43 different signers. There are 38,336 video samples in total. The samples are recorded using Microsoft Kinect v2 in RGB, depth and skel...
Provide a detailed description of the following dataset: AUTSL
WikiHowQA
WikiHowQA is a Community-based Question Answering dataset, which can be used for both answer selection and abstractive summarization tasks. It contains 76,687 questions in the train set, 8,000 in the development set and 22,354 in the test set.
Provide a detailed description of the following dataset: WikiHowQA
DRealSR
DRealSR establishes a Super Resolution (SR) benchmark with diverse real-world degradation processes, mitigating the limitations of conventional simulated image degradation. It has been collected from five DSLR cameras in natural scenes and cover indoor and outdoor scenes avoiding moving objects, e.g., advertising p...
Provide a detailed description of the following dataset: DRealSR
EDEN
EDEN (Enclosed garDEN) is a multimodal synthetic dataset, a dataset for nature-oriented applications. The dataset features more than 300K images captured from more than 100 garden models. Each image is annotated with various low/high-level vision modalities, including semantic segmentation, depth, surface normals, intr...
Provide a detailed description of the following dataset: EDEN
Wiki-CS
Wiki-CS is a Wikipedia-based dataset for benchmarking Graph Neural Networks. The dataset is constructed from Wikipedia categories, specifically 10 classes corresponding to branches of computer science, with very high connectivity. The node features are derived from the text of the corresponding articles. They were cal...
Provide a detailed description of the following dataset: Wiki-CS
ChaosNLI
Chaos NLI is a Natural Language Inference (NLI) dataset with 100 annotations per example (for a total of 464,500 annotations) for some existing data points in the development sets of SNLI, MNLI, and Abductive NLI. The dataset provides additional labels for NLI annotations that reflect the distribution of human annotato...
Provide a detailed description of the following dataset: ChaosNLI
VideoForensicsHQ
VideoForensicsHQ is a benchmark dataset for face video forgery detection, providing high quality visual manipulations. It is one of the first face video manipulation benchmark sets that also contains audio and thus complements existing datasets along a new challenging dimension. VideoForensicsHQ shows manipulations at...
Provide a detailed description of the following dataset: VideoForensicsHQ
SOLO
The SOLO Corpus comprises over 4 million English tweets, each of which contains at least one of the following tokens: solitude, lonely, and loneliness. The corpus has been collected to analyze the language and emotions associated with the state of being alone in English tweets. Tweets related to the state of being a...
Provide a detailed description of the following dataset: SOLO
EXPO-HD
The EXPO-HD Dataset is a dataset of Expo whiteboard markers for the purpose of instance segmentation. The dataset contains two subsets (both include instances segmentation labels): * Photorealistic synthetic image dataset with 5000 images. * Real image dataset with 200 images (used for validation and test). The ...
Provide a detailed description of the following dataset: EXPO-HD
Twitter Death Hoaxes
This is a dataset for detection fake death hoaxes. It consists of of death reports collected from Twitter between 1st January, 2012 and 31st December, 2014. It was collected by tracking the keyword 'RIP', and matching those tweets in which a name is mentioned next to RIP. Matching names were identified by using Wikidat...
Provide a detailed description of the following dataset: Twitter Death Hoaxes
KACC
The KACC benchmark consists of three subtasks that can be applied to knowledge graphs: knowledge abstraction, knowledge concretization and knowledge completion. - The **knowledge abstraction** subtask contains tasks of concept inference, schema prediction and concept graph completion on the two-view KG. - The **k...
Provide a detailed description of the following dataset: KACC
MSSD
The Spotify Music Streaming Sessions Dataset (MSSD) consists of 160 million streaming sessions with associated user interactions, audio features and metadata describing the tracks streamed during the sessions, and snapshots of the playlists listened to during the sessions. This dataset enables research on important...
Provide a detailed description of the following dataset: MSSD
InterHand2.6M
The InterHand2.6M dataset is a large-scale real-captured dataset with accurate GT 3D interacting hand poses, used for 3D hand pose estimation The dataset contains 2.6M labeled single and interacting hand frames.
Provide a detailed description of the following dataset: InterHand2.6M
TaxiNLI
TaxiNLI is a dataset collected based on the principles and categorizations of the aforementioned taxonomy. A subset of examples are curated from MultiNLI (Williams et al., 2018) by sampling uniformly based on the entailment label and the domain. The dataset is annotated with finegrained category labels.
Provide a detailed description of the following dataset: TaxiNLI
AllMusic Mood Subset
The AllMusic Mood Subset (AMS) is a dataset for mood classification from songs. It is created by matching a subset of the Million Song Dataset (MSD), totalling 67k tracks, with expert annotations of 188 different moods collected from AllMusic. Since the AMS is a subset of the MSD, the audio data is gathered by obtai...
Provide a detailed description of the following dataset: AllMusic Mood Subset
NISP
This dataset contains speech recordings along with speaker physical parameters (height, weight, shoulder size, age ) as well as regional information and linguistic information. There are a total of 345 speakers (219 male and 126 female). The dataset contains sentences that are taken out from newspapers. Each speaker...
Provide a detailed description of the following dataset: NISP
EDUVSUM
EDUVSUM contains educational videos with subtitles from three popular e-learning platforms: Edx,YouTube, and TIB AV-Portal that cover the following topics: crash course on history of science and engineering, computer science, python and web programming, machine learning and computer vision, Internet of things (IoT), an...
Provide a detailed description of the following dataset: EDUVSUM
ADVANCE
The AuDio Visual Aerial sceNe reCognition datasEt (ADVANCE) is a brand-new multimodal learning dataset, which aims to explore the contribution of both audio and conventional visual messages to scene recognition. This dataset in summary contains 5075 pairs of geotagged aerial images and sounds, classified into 13 scene ...
Provide a detailed description of the following dataset: ADVANCE
Multi-Modal CelebA-HQ
Multi-Modal-CelebA-HQ is a large-scale face image dataset that has 30,000 high-resolution face images selected from the CelebA dataset by following CelebA-HQ. Each image has high-quality segmentation mask, sketch, descriptive text, and image with transparent background. Multi-Modal-CelebA-HQ can be used to train and...
Provide a detailed description of the following dataset: Multi-Modal CelebA-HQ
Short Text Font Dataset
The proposed dataset includes 1,309 short text instances from Adobe Spark. The dataset is a collection of publicly available sample texts created by different designers. It covers a variety of topics found in posters, flyers, motivational quotes and advertisements.
Provide a detailed description of the following dataset: Short Text Font Dataset
SPHERE-calorie
The dataset contains both RGB and depth images, and the data from two accelerometers, together with ground truth calorie values from a calorimeter for calorie expenditure estimation in home environments.
Provide a detailed description of the following dataset: SPHERE-calorie
SmartCity
SmartCity consists of 50 images in total collected from ten city scenes including office entrance, sidewalk, atrium, shopping mall etc.. Unlike the existing crowd counting datasets with images of hundreds/thousands of pedestrians and nearly all the images being taken outdoors, SmartCity has few pedestrians in images an...
Provide a detailed description of the following dataset: SmartCity
ErhuPT
This dataset is an audio dataset containing about 1500 audio clips recorded by multiple professional players.
Provide a detailed description of the following dataset: ErhuPT
VideoNavQA
The VideoNavQA dataset contains pairs of questions and videos generated in the House3D environment. The goal of this dataset is to assess question-answering performance from nearly-ideal navigation paths, while considering a much more complete variety of questions than current instantiations of the Embodied Question An...
Provide a detailed description of the following dataset: VideoNavQA