dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
Pan+ChiPhoto
**Pan+ChiPhoto** dataset is a Chinese character dataset. It is built by the combination of two datasets: ChiPhoto and Pan_Chinese_Character dataset. The images in this dataset are mainly captured at outdoors in Beijing and Shanghai, China, which involve various scenes like signs, boards, advertisements, banners, object...
Provide a detailed description of the following dataset: Pan+ChiPhoto
ISI_Bengali_Character
The **ISI_Bengali_Character** dataset contains 158 classes of Bengali numerals, characters or their parts. 19,530 Bengali character samples are available. Most of the images in the dataset are synthesized. Source: [Boosting Scene Character Recognition by Learning Canonical Forms of Glyphs](https://arxiv.org/abs/1907.0...
Provide a detailed description of the following dataset: ISI_Bengali_Character
Florentine
The **Florentine** dataset is a dataset of facial gestures which contains facial clips from 160 subjects (both male and female), where gestures were artificially generated according to a specific request, or genuinely given due to a shown stimulus. 1032 clips were captured for posed expressions and 1745 clips for induc...
Provide a detailed description of the following dataset: Florentine
INRIA DLFD
The **INRIA Dense Light Field** Dataset (DLFD) is a dataset for testing depth estimation methods in a light field. DLFD contains 39 scenes with disparity range [-4,4] pixels. The light fields are of spatial resolution 512 x 512 and angular resolution 9 x 9. Source: [http://clim.inria.fr/Datasets/InriaSynLF/index.html]...
Provide a detailed description of the following dataset: INRIA DLFD
INRIA SLFD
The INRIA Sprse Light Field Dataset (SLFD) is a dataset for testing depth estimation methods in a light field. SLFD contains 53 scenes with disparity range [-20,20] pixels. The light fields are of spatial resolution 512 x 512 and angular resolution 9 x 9. Source: [http://clim.inria.fr/Datasets/InriaSynLF/index.html](h...
Provide a detailed description of the following dataset: INRIA SLFD
AIDS Antiviral Screen
The **AIDS Antiviral Screen** dataset is a dataset of screens checking tens of thousands of compounds for evidence of anti-HIV activity. The available screen results are chemical graph-structured data of these various compounds.
Provide a detailed description of the following dataset: AIDS Antiviral Screen
Retinal Microsurgery
The **Retinal Microsurgery** dataset is a dataset for surgical instrument tracking. It consists of 18 in-vivo sequences, each with 200 frames of resolution 1920 × 1080 pixels. The dataset is further classified into four instrument-dependent subsets. The annotated tool joints are n=3 and semantic classes c=2 (tool and b...
Provide a detailed description of the following dataset: Retinal Microsurgery
Daimler Monocular Pedestrian Detection
The **Daimler Monocular Pedestrian Detection** dataset is a dataset for pedestrian detection in urban environments. The training set contains 15560 pedestrian samples (image cut-outs at 48×96 resolution) and 6744 additional full images without pedestrians for extracting negative samples. The test set contains an indepe...
Provide a detailed description of the following dataset: Daimler Monocular Pedestrian Detection
ETHZ-Shape
The ETHZ Shape dataset contains images of five diverse shape-based classes, collected from Flickr and Google Images. The main challenges it offers are clutter, intra-class shape variability, and scale changes. The authors deliberately selected several images where the object comprises only a rather small portion of the...
Provide a detailed description of the following dataset: ETHZ-Shape
L-Bird
The **L-Bird** (**Large-Bird**) dataset contains nearly 4.8 million images which are obtained by searching images of a total of 10,982 bird species from the Internet.
Provide a detailed description of the following dataset: L-Bird
Extended BBC Pose
**Extended BBC Pose** is a pose estimation dataset which extends the BBC Pose dataset with 72 additional training videos. Combined with the original BBC TV dataset, the dataset contains 92 videos (82 training, 5 validation and 5 testing), i.e. around 7 million frames. The frames of the new 72 videos are automatically a...
Provide a detailed description of the following dataset: Extended BBC Pose
Short BBC Pose
**Short BBC Pose** contains five one-hour-long videos with sign language signers each with different sleeve length (in contrast to the BBC pose and Extended BBC Pose, which only contain signers with moderately long sleeves). Each of the five videos has 200 test frames (which have been manually annotated with joint loca...
Provide a detailed description of the following dataset: Short BBC Pose
ChaLearn Pose
**ChaLearn Pose** is a subset of the ChaLearn 2013 Multi-modal gesture dataset from Escalera et al. ICMI'13, which contains 23 hours of Kinect data of 27 persons performing 20 Italian gestures. The data includes RGB, depth, foreground segmentations and full body skeletons. In this dataset, both the training and testing...
Provide a detailed description of the following dataset: ChaLearn Pose
VoxCeleb2
**VoxCeleb2** is a large scale speaker recognition dataset obtained automatically from open-source media. VoxCeleb2 consists of over a million utterances from over 6k speakers. Since the dataset is collected ‘in the wild’, the speech segments are corrupted with real world noise including laughter, cross-talk, channel e...
Provide a detailed description of the following dataset: VoxCeleb2
VCTK
This CSTR **VCTK** Corpus includes speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent archive. The newspaper texts were taken from Herald Glasgow, ...
Provide a detailed description of the following dataset: VCTK
DIRHA
**DIRHA**-English is a multi-microphone database composed of real and simulated sequences of 1-minute. The overall corpus is composed of different types of sequences including: 1) Phonetically-rich sentences; 2) WSJ 5-k utterances; 3) WSJ 20-k utterances; 4) Conversational speech (also including keywords and commands)....
Provide a detailed description of the following dataset: DIRHA
VoxForge
**VoxForge** is an open speech dataset that was set up to collect transcribed speech for use with Free and Open Source Speech Recognition Engines (on Linux, Windows and Mac). Image Source: [http://www.voxforge.org/home](http://www.voxforge.org/home)
Provide a detailed description of the following dataset: VoxForge
Penn Action
The **Penn Action** Dataset contains 2326 video sequences of 15 different actions and human joint annotations for each sequence.
Provide a detailed description of the following dataset: Penn Action
FLIC
The **FLIC** dataset contains 5003 images from popular Hollywood movies. The images were obtained by running a state-of-the-art person detector on every tenth frame of 30 movies. People detected with high confidence (roughly 20K candidates) were then sent to the crowdsourcing marketplace Amazon Mechanical Turk to obtai...
Provide a detailed description of the following dataset: FLIC
WikiArt
**WikiArt** contains painting from 195 different artists. The dataset has 42129 images for training and 10628 images for testing.
Provide a detailed description of the following dataset: WikiArt
Sim10k
SIM10k is a synthetic dataset containing 10,000 images, which is rendered from the video game Grand Theft Auto V (GTA5).
Provide a detailed description of the following dataset: Sim10k
EYEDIAP
The **EYEDIAP** dataset is a dataset for gaze estimation from remote RGB, and RGB-D (standard vision and depth), cameras. The recording methodology was designed by systematically including, and isolating, most of the variables which affect the remote gaze estimation algorithms: * Head pose variations. * Person vari...
Provide a detailed description of the following dataset: EYEDIAP
G3D
The Gaming 3D Dataset (**G3D**) focuses on real-time action recognition in a gaming scenario. It contains 10 subjects performing 20 gaming actions: “punch right”, “punch left”, “kick right”, “kick left”, “defend”, “golf swing”, “tennis swing forehand”, “tennis swing backhand”, “tennis serve”, “throw bowling ball”, “aim...
Provide a detailed description of the following dataset: G3D
O-HAZE
The O-Haze dataset contains 35 hazy images (size 2833×4657 pixels) for training. It has 5 hazy images for validation along with their corresponding ground truth images.
Provide a detailed description of the following dataset: O-HAZE
UMIST
The Sheffield (previously **UMIST**) Face Database consists of 564 images of 20 individuals (mixed race/gender/appearance). Each individual is shown in a range of poses from profile to frontal views – each in a separate directory labelled 1a, 1b, … 1t and images are numbered consecutively as they were taken. The files ...
Provide a detailed description of the following dataset: UMIST
CVUSA
The CVUSA dataset is a matching task between street- and aerial views, from different regions of the US. This task helps to determine localization without GPS coordinates for the street-view images. Google Street View panoramas are used as ground images, and matching aerial images at zoom level 19 are obtained from Mic...
Provide a detailed description of the following dataset: CVUSA
FC100
The **FC100** dataset (**Fewshot-CIFAR100**) is a newly split dataset based on CIFAR-100 for few-shot learning. It contains 20 high-level categories which are divided into 12, 4, 4 categories for training, validation and test. There are 60, 20, 20 low-level classes in the corresponding split containing 600 images of si...
Provide a detailed description of the following dataset: FC100
PASCAL-5i
**PASCAL-5i** is a dataset used to evaluate few-shot segmentation. The dataset is sub-divided into 4 folds each containing 5 classes. A fold contains labelled samples from 5 classes that are used for evaluating the few-shot learning method. The rest 15 classes are used for training.
Provide a detailed description of the following dataset: PASCAL-5i
TrajNet
The **TrajNet** Challenge represents a large multi-scenario forecasting benchmark. The challenge consists on predicting 3161 human trajectories, observing for each trajectory 8 consecutive ground-truth values (3.2 seconds) i.e., t−7,t−6,…,t, in world plane coordinates (the so-called world plane Human-Human protocol) a...
Provide a detailed description of the following dataset: TrajNet
Set12
**Set12** is a collection of 12 grayscale images of different scenes that are widely used for evaluation of image denoising methods. The size of each image is 256×256.
Provide a detailed description of the following dataset: Set12
TotalCapture
The **TotalCapture** dataset consists of 5 subjects performing several activities such as walking, acting, a range of motion sequence (ROM) and freestyle motions, which are recorded using 8 calibrated, static HD RGB cameras and 13 IMUs attached to head, sternum, waist, upper arms, lower arms, upper legs, lower legs and...
Provide a detailed description of the following dataset: TotalCapture
I-HAZE
The I-Haze dataset contains 25 indoor hazy images (size 2833×4657 pixels) training. It has 5 hazy images for validation along with their corresponding ground truth images. Source: [Single image dehazing for a variety of haze scenarios using back projected pyramid network](https://arxiv.org/abs/2008.06713) Image Source...
Provide a detailed description of the following dataset: I-HAZE
SEED
The **SEED** dataset contains subjects' EEG signals when they were watching films clips. The film clips are carefully selected so as to induce different types of emotion, which are positive, negative, and neutral ones.
Provide a detailed description of the following dataset: SEED
SHREC
The **SHREC** dataset contains 14 dynamic gestures performed by 28 participants (all participants are right handed) and captured by the Intel RealSense short range depth camera. Each gesture is performed between 1 and 10 times by each participant in two way: using one finger and the whole hand. Therefore, the dataset i...
Provide a detailed description of the following dataset: SHREC
Florence3D
The dataset collected at the University of Florence during 2012, has been captured using a Kinect camera. It includes 9 activities: wave, drink from a bottle, answer phone,clap, tight lace, sit down, stand up, read watch, bow. During acquisition, 10 subjects were asked to perform the above actions for 2/3 times. This r...
Provide a detailed description of the following dataset: Florence3D
SNAP
**SNAP** is a collection of large network datasets. It includes graphs representing social networks, citation networks, web graphs, online communities, online reviews and more. [Social networks](http://snap.stanford.edu/data/#socnets) : online social networks, edges represent interactions between people [Networks...
Provide a detailed description of the following dataset: SNAP
BioASQ
**BioASQ** is a question answering dataset. Instances in the BioASQ dataset are composed of a question (Q), human-annotated answers (A), and the relevant contexts (C) (also called snippets).
Provide a detailed description of the following dataset: BioASQ
STRING
**STRING** is a collection of protein-protein interaction (PPI) networks.
Provide a detailed description of the following dataset: STRING
OpenWebText
**OpenWebText** is an open-source recreation of the [WebText](/dataset/webtext) corpus. The text is web content extracted from URLs shared on Reddit with at least three upvotes. (38GB).
Provide a detailed description of the following dataset: OpenWebText
Foursquare
The **Foursquare** dataset consists of check-in data for different cities. One subset contains check-ins in NYC and Tokyo collected for about 10 month (from 12 April 2012 to 16 February 2013). It contains 227,428 check-ins in New York city and 573,703 check-ins in Tokyo. Each check-in is associated with its time stamp,...
Provide a detailed description of the following dataset: Foursquare
PeerRead
PearRead is a dataset of scientific peer reviews. The dataset consists of over 14K paper drafts and the corresponding accept/reject decisions in top-tier venues including ACL, NIPS and ICLR, as well as over 10K textual peer reviews written by experts for a subset of the papers.
Provide a detailed description of the following dataset: PeerRead
Kinship
This relational database consists of 24 unique names in two families (they have equivalent structures). Source: [https://archive.ics.uci.edu/ml/datasets/kinship](https://archive.ics.uci.edu/ml/datasets/kinship)
Provide a detailed description of the following dataset: Kinship
Mindboggle
**Mindboggle** is a large publicly available dataset of manually labeled brain MRI. It consists of 101 subjects collected from different sites, with cortical meshes varying from 102K to 185K vertices. Each brain surface contains 25 or 31 manually labeled parcels.
Provide a detailed description of the following dataset: Mindboggle
Learning to Rank Challenge
The Yahoo! **Learning to Rank Challenge** dataset consists of 709,877 documents encoded in 700 features and sampled from query logs of the Yahoo! search engine, spanning 29,921 queries.
Provide a detailed description of the following dataset: Learning to Rank Challenge
Linux
The LINUX dataset consists of 48,747 Program Dependence Graphs (PDG) generated from the **Linux** kernel. Each graph represents a function, where a node represents one statement and an edge represents the dependency between the two statements Source: [Convolutional Set Matching for Graph Similarity](https://arxiv.org/...
Provide a detailed description of the following dataset: Linux
AMiner
The **AMiner** Dataset is a collection of different relational datasets. It consists of a set of relational networks such as citation networks, academic social networks or topic-paper-autor networks among others.
Provide a detailed description of the following dataset: AMiner
Email-EU
EmailEU is a directed temporal network constructed from email exchanges in a large European research institution for a 803-day period. It contains 986 email addresses as nodes and 332,334 emails as edges with timestamps. There are 42 ground truth departments in the dataset. Source: [gl2vec: Learning Feature Representa...
Provide a detailed description of the following dataset: Email-EU
IMDB-BINARY
**IMDB-BINARY** is a movie collaboration dataset that consists of the ego-networks of 1,000 actors/actresses who played roles in movies in IMDB. In each graph, nodes represent actors/actress, and there is an edge between them if they appear in the same movie. These graphs are derived from the Action and Romance genres.
Provide a detailed description of the following dataset: IMDB-BINARY
NCBI Disease
The **NCBI Disease** corpus consists of 793 PubMed abstracts, which are separated into training (593), development (100) and test (100) subsets. The NCBI Disease corpus is annotated with disease mentions, using concept identifiers from either MeSH or OMIM.
Provide a detailed description of the following dataset: NCBI Disease
arXiv Astro-Ph
Arxiv ASTRO-PH (Astro Physics) collaboration network is from the e-print arXiv and covers scientific collaborations between authors papers submitted to Astro Physics category. If an author i co-authored a paper with author j, the graph contains a undirected edge from i to j. If the paper is co-authored by k authors thi...
Provide a detailed description of the following dataset: arXiv Astro-Ph
MSLR-WEB10K
The **MSLR-WEB10K** dataset consists of 10,000 search queries over the documents from search results. The data also contains the values of 136 features and a corresponding user-labeled relevance factor on a scale of one to five with respect to each query-document pair. It is a subset of the MSLR-WEB30K dataset.
Provide a detailed description of the following dataset: MSLR-WEB10K
BeerAdvocate
BeerAdvocate is a dataset that consists of beer reviews from beeradvocate. The data span a period of more than 10 years, including all ~1.5 million reviews up to November 2011. Each review includes ratings in terms of five "aspects": appearance, aroma, palate, taste, and overall impression. Reviews include product and ...
Provide a detailed description of the following dataset: BeerAdvocate
Epinion
The **Epinion**s dataset is trust network dataset. For each user, it contains his profile, his ratings and his trust relations. For each rating, it has the product name and its category, the rating score, the time point when the rating is created, and the helpfulness of this rating. Source: [https://www.cse.msu.edu/~t...
Provide a detailed description of the following dataset: Epinion
Stanford Light Field
The **Stanford Light Field** Archive is a collection of several light fields for research in computer graphics and vision.
Provide a detailed description of the following dataset: Stanford Light Field
Arxiv GR-QC
**Arxiv GR-QC** (General Relativity and Quantum Cosmology) collaboration network is from the e-print arXiv and covers scientific collaborations between authors papers submitted to General Relativity and Quantum Cosmology category. If an author i co-authored a paper with author j, the graph contains a undirected edge fr...
Provide a detailed description of the following dataset: Arxiv GR-QC
Orkut
**Orkut** is a social network dataset consisting of friendship social network and ground-truth communities from Orkut.com on-line social network where users form friendship each other. Each connected component in a group is regarded as a separate ground-truth community. The ground-truth communities which have less t...
Provide a detailed description of the following dataset: Orkut
Friendster
**Friendster** is an on-line gaming network. Before re-launching as a game website, Friendster was a social networking site where users can form friendship edge each other. Friendster social network also allows users form a group which other members can then join. The Friendster dataset consist of ground-truth communit...
Provide a detailed description of the following dataset: Friendster
MQ2008
The **MQ2008** dataset is a dataset for Learning to Rank. It contains 800 queries with labelled documents.
Provide a detailed description of the following dataset: MQ2008
IMDB-MULTI
**IMDB-MULTI** is a relational dataset that consists of a network of 1000 actors or actresses who played roles in movies in IMDB. A node represents an actor or actress, and an edge connects two nodes when they appear in the same movie. In IMDB-MULTI, the edges are collected from three different genres: Comedy, Romance ...
Provide a detailed description of the following dataset: IMDB-MULTI
REDDIT-12K
Reddit12k contains 11929 graphs each corresponding to an online discussion thread where nodes represent users, and an edge represents the fact that one of the two users responded to the comment of the other user. There is 1 of 11 graph labels associated with each of these 11929 discussion graphs, representing the categ...
Provide a detailed description of the following dataset: REDDIT-12K
REDDIT-BINARY
**REDDIT-BINARY** consists of graphs corresponding to online discussions on Reddit. In each graph, nodes represent users, and there is an edge between them if at least one of them respond to the other’s comment. There are four popular subreddits, namely, IAmA, AskReddit, TrollXChromosomes, and atheism. IAmA and AskRedd...
Provide a detailed description of the following dataset: REDDIT-BINARY
MQ2007
The **MQ2007** dataset consists of queries, corresponding retrieved documents and labels provided by human experts. The possible relevance labels for each document are “relevant”, “partially relevant”, and “not relevant”.
Provide a detailed description of the following dataset: MQ2007
Amazon Fine Foods
Amazon Fine Foods is a dataset that consists of reviews of fine foods from amazon. The data span a period of more than 10 years, including all ~500,000 reviews up to October 2012. Reviews include product and user information, ratings, and a plaintext review.
Provide a detailed description of the following dataset: Amazon Fine Foods
REDDIT-5K
Reddit-5K is a relational dataset extracted from Reddit.
Provide a detailed description of the following dataset: REDDIT-5K
LastFM Asia
A social network of LastFM users which was collected from the public API in March 2020. Nodes are LastFM users from Asian countries and edges are mutual follower relationships between them. The vertex features are extracted based on the artists liked by the users. The task related to the graph is multinomial node class...
Provide a detailed description of the following dataset: LastFM Asia
EMNIST
**EMNIST** (extended MNIST) has 4 times more data than [MNIST](/dataset/mnist). It is a set of handwritten digits with a 28 x 28 format.
Provide a detailed description of the following dataset: EMNIST
Arcade Learning Environment
The **Arcade Learning Environment** (ALE) is an object-oriented framework that allows researchers to develop AI agents for Atari 2600 games. It is built on top of the Atari 2600 emulator Stella and separates the details of emulation from agent design.
Provide a detailed description of the following dataset: Arcade Learning Environment
MedleyDB
**MedleyDB**, is a dataset of annotated, royalty-free multitrack recordings. It was curated primarily to support research on melody extraction. For each song melody f₀ annotations are provided as well as instrument activations for evaluating automatic instrument recognition. The original dataset consists of 122 multitr...
Provide a detailed description of the following dataset: MedleyDB
MedleyDB 2.0
**MedleyDB 2.0** is a superset of the MedleyDB – a dataset of annotated, royalty-free multitrack recordings. The second iteration of the dataset includes 74 new multitrack recordings resulting in 194 songs in total. Source: [https://medleydb.weebly.com/](https://medleydb.weebly.com/) Image Source: [https://medleydb.we...
Provide a detailed description of the following dataset: MedleyDB 2.0
MIR-1K
**MIR-1K** (Multimedia Information Retrieval lab, 1000 song clips) is a dataset designed for singing voice separation. It contains: * 1000 song clips with the music accompaniment and the singing voice recorded as left and right channels, respectively, * Manual annotations of pitch contours in semitone, indices and ...
Provide a detailed description of the following dataset: MIR-1K
MagnaTagATune
**MagnaTagATune** dataset contains 25,863 music clips. Each clip is a 29-seconds-long excerpt belonging to one of the 5223 songs, 445 albums and 230 artists. The clips span a broad range of genres like Classical, New Age, Electronica, Rock, Pop, World, Jazz, Blues, Metal, Punk, and more. Each audio clip is supplied wit...
Provide a detailed description of the following dataset: MagnaTagATune
Lakh MIDI Dataset
The Lakh MIDI dataset is a collection of 176,581 unique MIDI files, 45,129 of which have been matched and aligned to entries in the Million Song Dataset. Its goal is to facilitate large-scale music information retrieval, both symbolic (using the MIDI files alone) and audio content-based (using information extracted fro...
Provide a detailed description of the following dataset: Lakh MIDI Dataset
iKala
The **iKala** dataset is a singing voice separation dataset that comprises of 252 30-second excerpts sampled from 206 iKala songs (plus 100 hidden excerpts reserved for MIREX data mining contest). The music accompaniment and the singing voice are recorded at the left and right channels respectively. Additionally, the h...
Provide a detailed description of the following dataset: iKala
CAL500
**CAL500** (**Computer Audition Lab 500**) is a dataset aimed for evaluation of music information retrieval systems. It consists of 502 songs picked from western popular music. The audio is represented as a time series of the first 13 Mel-frequency cepstral coefficients (and their first and second derivatives) extracte...
Provide a detailed description of the following dataset: CAL500
URMP
**URMP** (**University of Rochester Multi-Modal Musical Performance**) is a dataset for facilitating audio-visual analysis of musical performances. The dataset comprises 44 simple multi-instrument musical pieces assembled from coordinated but separately recorded performances of individual tracks. For each piece the dat...
Provide a detailed description of the following dataset: URMP
FMA
The **Free Music Archive** (**FMA**) is a large-scale dataset for evaluating several tasks in Music Information Retrieval. It consists of 343 days of audio from 106,574 tracks from 16,341 artists and 14,854 albums, arranged in a hierarchical taxonomy of 161 genres. It provides full-length and high-quality audio, pre-co...
Provide a detailed description of the following dataset: FMA
CCMixter
**CCMixter** is a singing voice separation dataset consisting of 50 full-length stereo tracks from [ccMixter](www.ccmixter.org) featuring many different musical genres. For each song there are three WAV files available: the background music, the voice signal, and their sum. Source: [Kernel Additive Models for Source S...
Provide a detailed description of the following dataset: CCMixter
GoodSounds
**GoodSounds** dataset contains around 28 hours of recordings of single notes and scales played by 15 different professional musicians, all of them holding a music degree and having some expertise in teaching. 12 different instruments (flute, cello, clarinet, trumpet, violin, alto sax alto, tenor sax, baritone sax, sop...
Provide a detailed description of the following dataset: GoodSounds
Jamendo Corpus
The **Jamendo Corpus** is a voice detection dataset consisting of 93 songs with Creative Commons license from the [Jamendo](http://www.jamendo.com/) free music sharing website. Segments of each song are annotated as “voice” (sung or spoken) or “no-voice”. The songs constitute a total of about 6 hours of music. The file...
Provide a detailed description of the following dataset: Jamendo Corpus
ForeDeCk
**ForeDeCk** is a time series database compiled at the National Technical University of Athens that contains 900,000 continuous time series, built from multiple, diverse and publicly accessible sources. ForeDeCk emphasizes business forecasting applications, including series from relevant domains such as industries, ser...
Provide a detailed description of the following dataset: ForeDeCk
M4
The **M4** dataset is a collection of 100,000 time series used for the fourth edition of the Makridakis forecasting Competition. The M4 dataset consists of time series of yearly, quarterly, monthly and other (weekly, daily and hourly) data, which are divided into training and test sets. The minimum numbers of observati...
Provide a detailed description of the following dataset: M4
MUSDB18-HQ
**MUSDB18-HQ** is a high-quality version of the MUSDB18 music tracks dataset. The high-quality dataset consists of the same 150 songs, but instead of MP4 files (compressed with Advanced Audio Coding encoder at 256kbps, with bandwidth limited to 16kHz), the songs are provided as raw WAV files. Image Source: [https://sig...
Provide a detailed description of the following dataset: MUSDB18-HQ
Slakh2100
The Synthesized Lakh (Slakh) Dataset is a dataset for audio source separation that is synthesized from the Lakh MIDI Dataset v0.1 using professional-grade sample-based virtual instruments. This first release of Slakh, called **Slakh2100**, contains 2100 automatically mixed tracks and accompanying MIDI files synthesized...
Provide a detailed description of the following dataset: Slakh2100
GuitarSet
**GuitarSet** is a dataset of high-quality guitar recordings and rich annotations. It contains 360 excerpts 30 seconds in length. The 360 excerpts are the result of the following combinations: * 6 players, * 2 versions: comping and soloing, * 5 styles: Rock, Singer-Songwriter, Bossa Nova, Jazz, and Funk, * 3 progressi...
Provide a detailed description of the following dataset: GuitarSet
Mixing Secrets
**Mixing Secrets** is an instrument recognition dataset containing 258 multi-track recordings sourced from the [Mixing Secrets for The Small Studio]( https://www.cambridge-mt.com/ms/mtk/) website. The dataset was labelled to be consistent with MedleyDB format. Source: [Mixing secrets: a multi-track dataset for instrum...
Provide a detailed description of the following dataset: Mixing Secrets
OpenMIC-2018
**OpenMIC-2018** is an instrument recognition dataset containing 20,000 examples of Creative Commons-licensed music available on the [Free Music Archive](http://freemusicarchive.org/). Each example is a 10-second excerpt which has been partially labeled for the presence or absence of 20 instrument classes by annotators...
Provide a detailed description of the following dataset: OpenMIC-2018
CAL500exp
The **CAL500 Expansion** (**CAL500exp**) dataset is an enriched version of the CAL500 music information retrieval dataset. CAL500exp is designed to facilitate music auto-tagging in a smaller temporal scale. The dataset consists of the same songs split into 3,223 acoustically homogenous segments of 3 to 16 seconds. The ...
Provide a detailed description of the following dataset: CAL500exp
CAL10K
The **CAL10K** dataset (introduced as Swat10k) contains 10,870 songs that are weakly-labelled using a tag vocabulary of 475 acoustic tags and 153 genre tags. The tags have all been harvested from [Pandora’s](https://www.pandora.com/) website and result from song annotations performed by expert musicologists involved wi...
Provide a detailed description of the following dataset: CAL10K
MuseScore
The **MuseScore** dataset is a collection of 344,166 audio and MIDI pairs downloaded from [MuseScore](https://musescore.org/) website. The audio is usually synthesized by the MuseScore synthesizer. The audio clips have diverse musical genres and are about two mins long on average. Due to copyright issues the dataset...
Provide a detailed description of the following dataset: MuseScore
MTG-Jamendo
The **MTG-Jamendo** dataset is an open dataset for music auto-tagging. The dataset contains over 55,000 full audio tracks with 195 tags categories (87 genre tags, 40 instrument tags, and 56 mood/theme tags). It is built using music available at Jamendo under Creative Commons licenses and tags provided by content upload...
Provide a detailed description of the following dataset: MTG-Jamendo
LibriCount
**LibriCount** is a synthetic dataset for speaker count estimation. The dataset contains a simulated cocktail party environment of 0 to 10 speakers, mixed with 0dB SNR from random utterances of different speakers from the LibriSpeech `CleanTest` dataset. All recordings are of 5s durations, and all speakers are active f...
Provide a detailed description of the following dataset: LibriCount
MultiWOZ
The **Multi-domain Wizard-of-Oz** (**MultiWOZ**) dataset is a large-scale human-human conversational corpus spanning over seven domains, containing 8438 multi-turn dialogues, with each dialogue averaging 14 turns. Different from existing standard datasets like WOZ and DSTC2, which contain less than 10 slots and only a ...
Provide a detailed description of the following dataset: MultiWOZ
ReVerb Challenge
The REVERB (**REverberant Voice Enhancement and Recognition Benchmark**) challenge is a benchmark for evaluation of automatic speech recognition techniques. The challenge assumes the scenario of capturing utterances spoken by a single stationary distant-talking speaker with 1-channe, 2-channel or 8-channel microphone-a...
Provide a detailed description of the following dataset: ReVerb Challenge
MPQA Opinion Corpus
The **MPQA Opinion Corpus** contains 535 news articles from a wide variety of news sources manually annotated for opinions and other private states (i.e., beliefs, emotions, sentiments, speculations, etc.).
Provide a detailed description of the following dataset: MPQA Opinion Corpus
DROP
**Discrete Reasoning Over Paragraphs** **DROP** is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a question, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or sorting). These operations require a m...
Provide a detailed description of the following dataset: DROP
New York Times Annotated Corpus
The **New York Times Annotated Corpus** contains over 1.8 million articles written and published by the New York Times between January 1, 1987 and June 19, 2007 with article metadata provided by the New York Times Newsroom, the New York Times Indexing Service and the online production staff at nytimes.com. The corpus i...
Provide a detailed description of the following dataset: New York Times Annotated Corpus
VisDial
**Visual Dialog** (**VisDial**) dataset contains human annotated questions based on images of MS COCO dataset. This dataset was developed by pairing two subjects on Amazon Mechanical Turk to chat about an image. One person was assigned the job of a ‘questioner’ and the other person acted as an ‘answerer’. The questione...
Provide a detailed description of the following dataset: VisDial
AMR Bank
The **AMR Bank** is a set of English sentences paired with simple, readable semantic representations. Version 3.0 released in 2020 consists of 59,255 sentences. Each AMR is a single rooted, directed graph. AMRs include PropBank semantic roles, within-sentence coreference, named entities and types, modality, negation...
Provide a detailed description of the following dataset: AMR Bank
WMT 2016
**WMT 2016** is a collection of datasets used in shared tasks of the First Conference on Machine Translation. The conference builds on ten previous Workshops on statistical Machine Translation. The conference featured ten shared tasks: - a news translation task, - an IT domain translation task, - a biomedical t...
Provide a detailed description of the following dataset: WMT 2016
WMT 2016 News
News translation is a recurring WMT task. The test set is a collection of parallel corpora consisting of about 1500 English sentences translated into 5 languages (Czech, German, Finnish, Romanian, Russian, Turkish) and additional 1500 sentences from each of the 5 languages translated to English. For Romanian a third of...
Provide a detailed description of the following dataset: WMT 2016 News