dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
DeepBeam
It contains 19 HDF5 files that represent a data collection campaign run on the NI mmWave Transceiver System with four SiBeam 60 GHz radio heads and on two Pi-Radio digital 60 GHz radios. Please refer to the website deepbeam.net
Provide a detailed description of the following dataset: DeepBeam
Amharic Error Corpus
Amharic Error Corpus is a manually annotated spelling error corpus for Amharic, lingua franca in Ethiopia. The corpus is designed to be used for the evaluation of spelling error detection and correction. The misspellings are tagged as non-word and real-word errors. In addition, the contextual information available in t...
Provide a detailed description of the following dataset: Amharic Error Corpus
BarkNet 1.0
23,000 cropped images of tree bark, for 23 species of trees around Quebec City, Canada. The images were captured at a distance between 20-60 cm away from the trunk. Labels include: individual tree ID, its species, and its DBH (diameter at breast height). Pictures were taken with four different devices: Nexus 5, Samsung...
Provide a detailed description of the following dataset: BarkNet 1.0
Epilepsy seizure prediction
The original dataset from the reference consists of 5 different folders, each with 100 files, with each file representing a single subject/person. Each file is a recording of brain activity for 23.6 seconds. The corresponding time-series is sampled into 4097 data points. Each data point is the value of the EEG recordin...
Provide a detailed description of the following dataset: Epilepsy seizure prediction
SymbolicData
This dataset is a collection of input-label pairs where each input is in the form of a numerical dataset, itself a set of input and output pairs {(x, y)}, and the corresponding label is a string encoding the symbolic expression governing the relationship between variables in the numerical dataset.
Provide a detailed description of the following dataset: SymbolicData
RadGraph
RadGraph is a dataset of entities and relations in radiology reports based on our novel information extraction schema, consisting of 600 reports with 30K radiologist annotations and 221K reports with 10.5M automatically generated annotations. We release a development dataset, which contains board-certified radiologist...
Provide a detailed description of the following dataset: RadGraph
CASIA-Iris-Complex
# Introduction Iris is considered one of the most accurate and reliable biometric modality. Iris is more stable and distinctive compared with fingerprint, face, voice, etc, and difficult to be replicated for spoof attacks. Although an iris pattern is naturally an ideal identifier, the development of a high-performan...
Provide a detailed description of the following dataset: CASIA-Iris-Complex
Extended YouTube Faces (E-YTF)
The proposed Extended-YouTube Faces (E-YTF) is an extension of the famous YouTube Faces (YTF) dataset and is specifically designed to further push the challenges of face recognition by addressing the problem of open-set face identification from heterogeneous data i.e. still images vs video.
Provide a detailed description of the following dataset: Extended YouTube Faces (E-YTF)
Amazon-PQA
**Amazon-PQA** is a product question-answer dataset. The Amazon-PQA dataset includes questions and their answers that are published on Amazon website, along with the public product information and category (Amazon Browse Node name). It contains more than 8M questions from 1M+ products.
Provide a detailed description of the following dataset: Amazon-PQA
SSL
This is a dataset to benchmark real-time embedded object detection models for RoboCup SSL (Small Size League).
Provide a detailed description of the following dataset: SSL
FilmStills
FilmStills is a dataset of stills taken from a variety of films and TV shows, each concatenated with a color-compressed (with a factor of 2.667) version of itself.
Provide a detailed description of the following dataset: FilmStills
LCO CR Dataset
Cosmic rays in the LCO CR dataset are labeled accurately and consistently across many diverse observations from various instruments. To the best of our knowledge, this is the largest dataset of its kind. It consists of over 4,500 scientific images from Las Cumbres Observatory global telescope network's 23 instruments. ...
Provide a detailed description of the following dataset: LCO CR Dataset
Message Content Rephrasing
We introduce a new task of rephrasing for amore natural virtual assistant. Currently, vir-tual assistants work in the paradigm of intent-slot tagging and the slot values are directlypassed as-is to the execution engine. However,this setup fails in some scenarios such as mes-saging when the query given by the user needs...
Provide a detailed description of the following dataset: Message Content Rephrasing
TAU-NIGENS Spatial Sound Events 2021
The TAU-NIGENS Spatial Sound Events 2021 dataset contains multiple spatial sound-scene recordings, consisting of sound events of distinct categories integrated into a variety of acoustical spaces, and from multiple source directions and distances as seen from the recording position. The spatialization of all sound even...
Provide a detailed description of the following dataset: TAU-NIGENS Spatial Sound Events 2021
PAD
**PAD** (Purpose-driven Affordance Dataset) is a dataset for affordance detection, which refers to identifying the potential action possibilities of objects in an image, which is an important ability for robot perception and manipulation. The dataset consists of 4K images from 31 affordance and 72 object categories.
Provide a detailed description of the following dataset: PAD
XL-Sum
**XL-Sum** is a comprehensive and diverse dataset for abstractive summarization comprising 1 million professionally annotated article-summary pairs from BBC, extracted using a set of carefully designed heuristics. The dataset covers 44 languages ranging from low to high-resource, for many of which no public dataset is ...
Provide a detailed description of the following dataset: XL-Sum
TNCR Dataset
We present TNCR, a new table dataset with varying image quality collected from free open source websites. TNCR dataset can be used for table detection in scanned document images and their classification into 5 different classes. TNCR contains 9428 high-quality labeled images. In this paper, we have implemented state...
Provide a detailed description of the following dataset: TNCR Dataset
HKR
The database is written in Cyrillic and shares the same 33 characters. Besides these characters, the Kazakh alphabet also contains 9 additional specific characters. This dataset is a collection of forms. The sources of all the forms in the datasets were generated by LATEX which subsequently was filled out by persons wi...
Provide a detailed description of the following dataset: HKR
HT Docking
**HT Docking** is a dataset consisting of 200 million 3D complex structures and 2D structure scores across a consistent set of 13 million ``in-stock'' molecules over 15 receptors, or binding sites, across the SARS-CoV-2 proteome. It is used to study surrogate model accuracy for protein-ligand docking.
Provide a detailed description of the following dataset: HT Docking
PointQA
**PointQA** is a set of datasets for Visual Question Datasets (VQA) that require a pointer to an object in the image to be answered correctly. The different datasets are: PointQA-Local, PointQA-LookTwice and PointQA-General.
Provide a detailed description of the following dataset: PointQA
TinyFace
**TinyFace** is a large scale face recognition benchmark to facilitate the investigation of natively LRFR (Low Resolution Face Recognition) at large scales (large gallery population sizes) in deep learning. The TinyFace dataset consists of 5,139 labelled facial identities given by 169,403 native LR face images (average...
Provide a detailed description of the following dataset: TinyFace
FB15K237-Refined
FB15K237-Refined is a refined version of FB15k237 by KGRefiner.
Provide a detailed description of the following dataset: FB15K237-Refined
WN18RR Refined
WN18RR Refined is a refined version of [WN18RR](https://paperswithcode.com/dataset/wn18rr) by KGRefiner
Provide a detailed description of the following dataset: WN18RR Refined
CHORD
CHORD is the first chorus recognition dataset containing 627 songs for public use.
Provide a detailed description of the following dataset: CHORD
TrajAir: A General Aviation Trajectory Dataset
This dataset contains aircraft trajectories in an untowered terminal airspace collected over 8 months surrounding the Pittsburgh-Butler Regional Airport [ICAO:KBTP], a single runway GA airport, 10 miles North of the city of Pittsburgh, Pennsylvania. The trajectory data is recorded using an on-site setup that includes a...
Provide a detailed description of the following dataset: TrajAir: A General Aviation Trajectory Dataset
PathQuestion
Adopts two subsets of Freebase (Bollacker et al., 2008) as Knowledge Bases to construct the PathQuestion (PQ) and the PathQuestion-Large (PQL) datasets. Paths are extracted between two entities which span two hops (es → r1 → e1 → r2 → a, denoted by -2H) or three hops (es→ r1 → e1 →r2 → e2→ r3 → a, denoted by -3H) and t...
Provide a detailed description of the following dataset: PathQuestion
PELD
PELD is a text-based emotional dialog dataset with personality traits for speakers. The dialogues in PELD are merged from the emotional dialogues in MELD and EmoryNLP , as well as the personality trait annotations from FriendsPersona. The personality traits in PELD are adopted from the personality annotations in ...
Provide a detailed description of the following dataset: PELD
VoxLingua107
VoxLingua107 is a dataset for spoken language recognition of 6628 hours (62 hours per language on the average) and it is accompanied by an evaluation set of 1609 verified utterances.
Provide a detailed description of the following dataset: VoxLingua107
MultiSubs
MultiSubs is a dataset of multilingual subtitles gathered from [the OPUS OpenSubtitles dataset](https://opus.nlpl.eu/OpenSubtitles.php), which in turn was sourced from [opensubtitles.org](http://www.opensubtitles.org/). We have supplemented some text fragments (visually salient nouns in this release) within the subtitl...
Provide a detailed description of the following dataset: MultiSubs
ISPRS Potsdam
The data set contains 38 patches (of the same size), each consisting of a true orthophoto (TOP) extracted from a larger TOP mosaic.
Provide a detailed description of the following dataset: ISPRS Potsdam
ISPRS Vaihingen
The data set contains 33 patches (of different sizes), each consisting of a true orthophoto (TOP) extracted from a larger TOP mosaic.
Provide a detailed description of the following dataset: ISPRS Vaihingen
BrazilDam Dataset
BrazilDAM is a multi sensor and multitemporal dataset that consists of multispectral images of ore tailings dams throughout Brazil. Landsat 8 and Sentinel 2 satellites that capture multispectral images over the years 2016, 2017, 2018 and 2019 were used. The dataset contains samples collected in different regions, which...
Provide a detailed description of the following dataset: BrazilDam Dataset
Brazilian Coffee Scenes Dataset
This dataset is a composition of scenes taken by SPOT sensor in 2005 over four counties in the State of Minas Gerais, Brazil: Arceburgo, Guaranesia, Guaxupé and Monte Santo. It has multispectral high-resolution scenes of coffee crops and non-coffee areas. It has many intraclass variance caused by different crop managem...
Provide a detailed description of the following dataset: Brazilian Coffee Scenes Dataset
SinGAN-Seg-polyps
**SinGAN-Seg-polyps** is a synthetic dataset for polyp segmentation consisting of 10,000 synthetic polyps and masks.
Provide a detailed description of the following dataset: SinGAN-Seg-polyps
XWINO
XWINO is a multilingual collection of Winograd Schemas in six languages that can be used for evaluation of cross-lingual commonsense reasoning capabilities. The datasets that comprise XWINO are: * Source: The original [Winograd Schema Challenge](https://cs.nyu.edu/~davise/papers/WinogradSchemas/WSCollection.xm...
Provide a detailed description of the following dataset: XWINO
NNE
**NNE** is a dataset for Nested Named Entity Recognition in English Newswire
Provide a detailed description of the following dataset: NNE
Antibody Watch
**Antibody Watch** is a dataset of text snippets extracted from over 2000 PubMed articles with annotations denoting specificity of antibodies.
Provide a detailed description of the following dataset: Antibody Watch
COVID-19 & Election
These datasets were used in the paper 'Evaluation of Thematic Coherence in Microblogs' (ACL, 2021). The data is structured as follows: each file represents a cluster of tweets which contains the tweet IDs, the journalist annotations for quality evaluation and issue identification, as well as the metric evaluation score...
Provide a detailed description of the following dataset: COVID-19 & Election
ZooScanNet
Plankton was sampled with various nets, from bottom or 500m depth to the surface, in many oceans of the world. Samples were imaged with a ZooScan. The full images were processed with ZooProcess which generated regions of interest (ROIs) around each individual object and a set of associated features measured on the obje...
Provide a detailed description of the following dataset: ZooScanNet
iMiGUE
**iMiGUE** is a dataset for emotional artificial intelligence research: identity-free video dataset for Micro-Gesture Understanding and Emotion analysis (iMiGUE). Different from existing public datasets, iMiGUE focuses on nonverbal body gestures without using any identity information, while the predominant researches o...
Provide a detailed description of the following dataset: iMiGUE
MultiCite
**MultiCite** is a dataset of 12,653 citation contexts from over 1,200 computational linguistics papers used for Citation context analysis (CCA). MultiCite contains multi-sentence, multi-label citation contexts within full paper texts.
Provide a detailed description of the following dataset: MultiCite
CityNet
**CityNet** is a multi-modal urban dataset containing data from 7 cities, each of which coming from 3 data sources, which can be used for urban computing and smart city research. The dataset consists of 3 types of raw data (city layout, taxi, meteorology) collected from 7 cities.
Provide a detailed description of the following dataset: CityNet
CrowdSpeech
**CrowdSpeech** is a publicly available large-scale dataset of crowdsourced audio transcriptions. It contains annotations for more than 20 hours of English speech from more than 1,000 crowd workers.
Provide a detailed description of the following dataset: CrowdSpeech
Toloka Business ID Recognition
This dataset, commissioned by the Yandex Business Directory, contains 10,000 photos of organization information signs shot in the Russian Federation along with the INN (taxpayer ID) and OGRN (Primary State Registration Number) codes shown on these signs. Toloka was used for both capturing photos and recognizing INN and...
Provide a detailed description of the following dataset: Toloka Business ID Recognition
pd4ml
**pd4ml** is a collection of datasets from fundamental physics research -- including particle physics, astroparticle physics, and hadron- and nuclear physics -- for supervised machine learning studies. These datasets, containing hadronic top quarks, cosmic-ray induced air showers, phase transitions in hadronic matter, ...
Provide a detailed description of the following dataset: pd4ml
Toloka WaterMeters
This datase, contains 1244 images of hot and cold water meters as well as their readings and coordinates of the displays showing those readings. Each image contains exactly one water meter. The archive also includes the pictures of the results of segmentation with the masks and collages. Toloka was used for photo captu...
Provide a detailed description of the following dataset: Toloka WaterMeters
RuADReCT
Created as part of the Social Media Mining for Health Applications (#SMM4H '20) shared tasks, this dataset consists of 9515 tweets describing health issues. Each tweet is labeled for whether it contains information about an adverse side effect that occurred when taking a drug. The dataset was a joint effort with the UP...
Provide a detailed description of the following dataset: RuADReCT
LRWC
This dataset contains the opinions of Russian native speakers about the relationship between a generic term (hypernym) and a specific instance of it (hyponym). Assembled by Dmitry Ustalov in 2017. A set of 300 most frequent nouns was extracted from the Russian National Corpus. Then each method or resource (including Ru...
Provide a detailed description of the following dataset: LRWC
Human-Annotated Sense-Disambiguated Word Contexts for Russian
This dataset contains human-annotated sense identifiers for 2562 contexts of 20 words used in the RUSSE'2018 shared task on Word Sense Induction and Disambiguation for the Russian language. Assembled by Dmitry Ustalov in 2017. In particular, 80 pre-annotated contexts were used for training the human annotators, and 256...
Provide a detailed description of the following dataset: Human-Annotated Sense-Disambiguated Word Contexts for Russian
ScanBank
ScanBank is a benchmark dataset for figure extraction from scanned electronic theses and dissertations containing 10 thousand scanned page images, manually labeled by humans as to the presence of the 3.3 thousand figures or tables found therein.
Provide a detailed description of the following dataset: ScanBank
Florence 3D actions dataset
The dataset collected at the University of Florence during 2012, has been captured using a Kinect camera. It includes 9 activities: wave, drink from a bottle, answer phone,clap, tight lace, sit down, stand up, read watch, bow. During acquisition, 10 subjects were asked to perform the above actions for 2/3 times. This r...
Provide a detailed description of the following dataset: Florence 3D actions dataset
Delaunay triangulation
Delaunay triangulation dataset for 5, 10, 15, 20 points. Both random and sorted datasets are included. If you have any trouble to use this dataset, contact hunnino10@gmail.com
Provide a detailed description of the following dataset: Delaunay triangulation
ExBAN
The ExBAN dataset: a corpus of NL explanations generated by crowd-sourced participants presented with the task of explaining simple Bayesian Network (BN) graphical representations. These explanations, in a separate collection effort, are rated for clarity and informativeness.
Provide a detailed description of the following dataset: ExBAN
ObMan-Ego
The ObMan-Ego is a large-scale synthetic hand dataset with egocentric scenes in which the simulated hands are provided by ObMan. The dataset is used for a hand segmentation task and its sim-to-real adaptation benchmark. Training, validation, and testing sets contain 150, 000, 6, 500, and 6, 500 images, respectively.
Provide a detailed description of the following dataset: ObMan-Ego
CPTC-2018
Intrusion alert dataset captured through the Collegiate Penetration Testing Competition (CPTC) 2018. Contains alerts from 6 student teams. For details, see "A Cybersecurity Dataset Derived from the National Collegiate Penetration Testing Competition" by Nathan Munaiah et al.
Provide a detailed description of the following dataset: CPTC-2018
SURREALvols
Added information about the subject's body height and volumes of 14 individual body parts.
Provide a detailed description of the following dataset: SURREALvols
Fingerprint Dataset
This dataset includes all music sources, background noises and impulse-reponses (IR) samples and conversation speech that have been used in the work "Neural Audio Fingerprint for High-specific Audio Retrieval based on Contrastive Learning" ICASSP 2021 (https://arxiv.org/abs/2010.11910).
Provide a detailed description of the following dataset: Fingerprint Dataset
Steel Tube Dataset
8 kinds of weld defects
Provide a detailed description of the following dataset: Steel Tube Dataset
Geography of Open Source Software
This dataset reports counts of active GitHub contributors (activity: 2019/2020) geolocated in early 2021. Counts are aggregated at the country level and at various regional scales. Besides countries, we report data on the EU NUTS2 level, for Brazilian, Russian, Chinese, Japanese, Indian, and US-American subnational geo...
Provide a detailed description of the following dataset: Geography of Open Source Software
PDE solutions
In this folder, you will find solutions of the following partial differential equations: - Burgers - Kortweg-de-Vries -Newell-Whitehead - Kuramoto-Sivashinsky You will find more info about how these were generated in the supplementary material of the paper: https://arxiv.org/abs/2106.11936
Provide a detailed description of the following dataset: PDE solutions
PDEs
In this dataset, you will find solutions of the following partial differential equations: - Burgers - Kortweg-de-Vries -Newell-Whitehead - Kuramoto-Sivashinsky You will find more info about how these were generated in the supplementary material of the paper: https://arxiv.org/abs/2106.11936
Provide a detailed description of the following dataset: PDEs
IowaRain
**IowaRain** is a dataset of rainfall events for the state of Iowa (2016-2019) acquired from the National Weather Service Next Generation Weather Radar (NEXRAD) system and processed by a quantitative precipitation estimation system. The dataset presented in this study could be used for better disaster monitoring, respo...
Provide a detailed description of the following dataset: IowaRain
Kosp2e
**Kosp2e** (read as `kospi'), is a corpus that allows Korean speech to be translated into English text in an end-to-end manner
Provide a detailed description of the following dataset: Kosp2e
HumanoidRobotPose
The **HumanoidRobotPose** dataset is a dataset for real-time pose estimation of humanoid robots.
Provide a detailed description of the following dataset: HumanoidRobotPose
FaVIQ
**FaVIQ** (Fact Verification from Information-seeking Questions) is a challenging and realistic fact verification dataset that reflects confusions raised by real users. We use the ambiguity in information-seeking questions and their disambiguation, and automatically convert them to true and false claims. These claims a...
Provide a detailed description of the following dataset: FaVIQ
OPA
Object-Placement-Assessment (OPA) is a task consisting on verifying whether a composite image is plausible in terms of the object placement. The foreground object should be placed at a reasonable location on the background considering location, size, occlusion, semantics, and etc. **OPA** is a synthesised dataset fo...
Provide a detailed description of the following dataset: OPA
DPPIN
**DPPIN** is a collection of dynamic networks, which consists of twelve generated dynamic protein-protein interaction networks of yeast cells, stored in twelve folders.
Provide a detailed description of the following dataset: DPPIN
MineRL BASALT
**MineRL BASALT** is an RL competition on solving human-judged tasks. The tasks in this competition do not have a pre-defined reward function: the goal is to produce trajectories that are judged by real humans to be effective at solving a given task.
Provide a detailed description of the following dataset: MineRL BASALT
SBU-WSD-Corpus
**SBU-WSD-Corpus** is a corpus for Persian Word Sense Disambiguation (WSD). It is manually annotated with senses from the Persian WordNet (FarsNet) sense inventory. SBU-WSD-Corpus consists of 19 Persian documents in different domains such as Sports, Science, Arts, etc. It includes 5892 content words of Persian running ...
Provide a detailed description of the following dataset: SBU-WSD-Corpus
VinDr-RibCXR
**VinDr-RibCXR** is a benchmark dataset for automatic segmentation and labeling of individual ribs from chest X-ray (CXR) scans. The VinDr-RibCXR contains 245 CXRs with corresponding ground truth annotations provided by human experts.
Provide a detailed description of the following dataset: VinDr-RibCXR
Disaster
**Disaster** is a dataset that contains images collected from various sources for three different disasters: fire, water and land. Besides this, it also contains images for various damaged infrastructure due to natural or man made calamities and damaged human due to war or accidents. There are 13,720 manually annota...
Provide a detailed description of the following dataset: Disaster
Google Landmarks
The **Google Landmarks** dataset contains 1,060,709 images from 12,894 landmarks, and 111,036 additional query images. The images in the dataset are captured at various locations in the world, and each image is associated with a GPS coordinate. This dataset is used to train and evaluate large-scale image retrieval mode...
Provide a detailed description of the following dataset: Google Landmarks
KiTS19
The 2021 Kidney and Kidney Tumor Segmentation challenge (abbreviated KiTS21) is a competition in which teams compete to develop the best system for automatic semantic segmentation of renal tumors and surrounding anatomy. [The 2021 Kidney and Kidney Tumor Segmentation Challenge](https://kits21.kits-challenge.org/) ...
Provide a detailed description of the following dataset: KiTS19
UrbanScene3D
UrbanScene3D is a large scale urban scene dataset associated with a handy simulator based on Unreal Engine 4 and AirSim, which consists of both man-made and real-world reconstruction scenes in different scales, referred to as UrbanScene3D. The manually made scene models have compact structures, which are carefully cons...
Provide a detailed description of the following dataset: UrbanScene3D
ChangeSim
**ChangeSim** is a dataset aimed at online scene change detection (SCD) and more. The data is collected in photo-realistic simulation environments with the presence of environmental non-targeted variations, such as air turbidity and light condition changes, as well as targeted object changes in industrial indoor enviro...
Provide a detailed description of the following dataset: ChangeSim
Red MiniImageNet 20% label noise
Part of the Controlled Noisy Web Labels Dataset.
Provide a detailed description of the following dataset: Red MiniImageNet 20% label noise
Red MiniImageNet 40% label noise
Part of the Controlled Noisy Web Labels Dataset.
Provide a detailed description of the following dataset: Red MiniImageNet 40% label noise
Red MiniImageNet 80% label noise
Part of the Controlled Noisy Web Labels Dataset.
Provide a detailed description of the following dataset: Red MiniImageNet 80% label noise
ISO17
### Description The molecules were randomly drawn from the largest set of isomers in the QM9 dataset [1] which consists of molecules with a fixed composition of atoms (C7O2H10) arranged in different chemically valid structures. It is an extension of the ismoer MD data used in [2]. The database was generated from ...
Provide a detailed description of the following dataset: ISO17
MCMD
A large-scale dataset in multi-programming languages and with rich information.
Provide a detailed description of the following dataset: MCMD
BCOPA-CE
We provide the BCOPA-CE test set, which has balanced token distribution in the correct and wrong alternatives and increases the difficulty of being aware of cause and effect. ### construction 1. for each premise of the 500 samples in COPA-test set, we generate one event manually which is a plausible answer to the op...
Provide a detailed description of the following dataset: BCOPA-CE
Multiple Testing and Variable Selection along Least Angle Regression's path
Data used in paper entitled "Multiple Testing and Variable Selection along Least Angle Regression's path". Zenodo file with the code of the paper arXiv:1906.12072
Provide a detailed description of the following dataset: Multiple Testing and Variable Selection along Least Angle Regression's path
HumanEval
This is an evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code". It used to measure functional correctness for synthesizing programs from docstrings. It consists of 164 original programming problems, assessing language comprehension, algo...
Provide a detailed description of the following dataset: HumanEval
Unbalance Classification Using Vibration Data
This dataset contains vibration data recorded on a rotating drive train. This drive train consists of an electronically commutated DC motor and a shaft driven by it, which passes through a roller bearing. With the help of a 3D-printed holder, unbalances with different weights and different radii were attached to the sh...
Provide a detailed description of the following dataset: Unbalance Classification Using Vibration Data
HYPE
HYPE Dataset - Version 1.0.0 REFERENCE PAPER ------------------- Morassi Sasso, A., Datta, S., Jeitler, M., Steckhan, N., Kessler, C. S., Michalsen, A., Arnrich, B., & Böttinger, E. (2020). HYPE: Predicting Blood Pressure from Photoplethysmograms in a Hypertensive Population. In M. Michalowski & R. Moskovitch (Eds...
Provide a detailed description of the following dataset: HYPE
Lakh Pianoroll Dataset
The Lakh Pianoroll Dataset (LPD) is a collection of 174,154 [multitrack pianorolls](https://salu133445.github.io/lakh-pianoroll-dataset/representation) derived from the [Lakh MIDI Dataset](http://colinraffel.com/projects/lmd/) (LMD). ## Getting the dataset We provide multiple subsets and versions of the dataset (see ...
Provide a detailed description of the following dataset: Lakh Pianoroll Dataset
Common Crawl
The Common Crawl corpus contains petabytes of data collected over 12 years of web crawling. The corpus contains raw web page data, metadata extracts and text extracts. Common Crawl data is stored on Amazon Web Services’ Public Data Sets and on multiple academic cloud platforms across the world.
Provide a detailed description of the following dataset: Common Crawl
CADSketchNet
CADSketchNet is an annotated collection of sketches of 3D CAD models. Dataset-A has 58,696 computer-generated sketches of the 3D CAD models across 68 categories of MCB. Dataset-B has 801 hand-drawn sketches of the 3D CAD models across 42 categories of ESB
Provide a detailed description of the following dataset: CADSketchNet
AIP Environment
AI Playground (AIP) is an open-source, Unreal Engine-based tool for generating and labeling virtual image data. With AIP, it is trivial to capture the same image under different conditions (e.g., fidelity, lighting, etc.) and with different ground truths (e.g., depth or surface normal values). AIP is easily extendable ...
Provide a detailed description of the following dataset: AIP Environment
Voice Conversion Challenge 2018
Voice conversion (VC) is a technique to transform a speaker identity included in a source speech waveform into a different one while preserving linguistic information of the source speech waveform. The Voice Conversion Challenge (VCC) 2016 was launched in 2016 at Interspeech 2016. The objective of the 2016 challenge wa...
Provide a detailed description of the following dataset: Voice Conversion Challenge 2018
SECBENCH
Dataset of 676 security vulnerabilities patches. In 2017, we mined the commits messages of 238 projects using regular expressions for each vulnerability (cf. Patterns). In 2020, we classified vulnerabilities using the CWE taxonomy. Some vulnerabilities contain the score and severity information (CVEs).
Provide a detailed description of the following dataset: SECBENCH
Sims4Action
* **The Sims4Action Dataset**: a videogame-based dataset for Synthetic→Real domain adaptation for human activity recognition. * **Goal** : Exploring the concept of constructing training examples for Activities of Daily Living (ADL) recognition by playing life simulation video games. * ** ***Sims4Action* dataset*...
Provide a detailed description of the following dataset: Sims4Action
GLIB: image dataset
data/images: data/images/Base : 132 screenshots of game1 & game2 with UI display issues from 466 test reports. data/images/Code : 9,412 screenshots of game1 & game2 with UI display issues generated by our Code augmentation method. data/images/Normal: 7,750 screenshots of game1 & game2 without UI display issues collect...
Provide a detailed description of the following dataset: GLIB: image dataset
Narvik Road Dataset
DIT4BEARs Internship Project (at UiT-The Arctic University of Norway) Dataset The dataset contains data of 5 months including weather conditions, friction coefficient, distance traveled, wind speed, surface temperature, air temperature, etc. This dataset was provided by DIT4BEARs for the Smart Road Internship Pro...
Provide a detailed description of the following dataset: Narvik Road Dataset
PackIt
The ability to jointly understand the geometry of objects and plan actions for manipulating them is crucial for intelligent agents. This ability is referred to as geometric planning. Recently, many interactive environments have been proposed to evaluate intelligent agents on various skills, however, none of them cater ...
Provide a detailed description of the following dataset: PackIt
EasyCom
The Easy Communications (EasyCom) dataset is a world-first dataset designed to help mitigate the cocktail party effect from an augmented-reality (AR) -motivated multi-sensor egocentric world view. The dataset contains AR glasses egocentric multi-channel microphone array audio, wide field-of-view RGB video, speech sourc...
Provide a detailed description of the following dataset: EasyCom
SportSett
This resource is designed to allow for research into Natural Language Generation. In particular, with neural data-to-text approaches although it is not limited to these.
Provide a detailed description of the following dataset: SportSett
NucMM
**NucMM** is a dataset for segmenting 3D cell nuclei from microscopy image volumes that pushes the task forward to the sub-cubic millimeter scale. It consists of two fully annotated volumes: one electron microscopy (EM) volume containing nearly the entire zebrafish brain with around 170,000 nuclei; and one micro-CT (uC...
Provide a detailed description of the following dataset: NucMM
AxonEM
The **AxonEM** dataset consists of two 30x30x30 um^3 EM image volumes from the human and mouse cortex, respectively. It is used for 3D axon instance segmentation of brain cortical regions. The authors proofread over 18,000 axon instances to provide dense 3D axon instance segmentation, enabling large-scale evaluation of...
Provide a detailed description of the following dataset: AxonEM
MSJudge
This is a challenging dataset from real courtrooms to predict the legal judgment in a reasonably encyclopedic manner by leveraging the genuine input of the case -- plaintiff's claims and court debate data, from which the case's facts are automatically recognized by comprehensively understanding the multi-role dialogues...
Provide a detailed description of the following dataset: MSJudge