dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
WebFace260M
**WebFace260M** is a million-scale face benchmark, which is constructed for the research community towards closing the data gap behind the industry. It consists of: - Noisy 4M identities and 260M faces - High-quality training data with 42M images of 2M identities by using automatic cleaning - A test set with rich...
Provide a detailed description of the following dataset: WebFace260M
DeepFake MNIST+
DeepFake MNIST+ is a deepfake facial animation dataset. The dataset is generated by a SOTA image animation generator. It includes 10,000 facial animation videos in ten different actions, which can spoof the recent liveness detectors.
Provide a detailed description of the following dataset: DeepFake MNIST+
mvor
**Multi-View Operating Room** (MVOR) dataset consists of 732 synchronized multi-view frames recorded by three RGB-D cameras in a hybrid OR during real clinical interventions. Each multi-view frame consists of three color and three depth images. The MVOR dataset was sampled from four days of recording in an intervention...
Provide a detailed description of the following dataset: mvor
ConvRef
**ConvRef** is a conversational QA benchmark with reformulations. It consists of around 11k natural conversations with about 205k reformulations. ConvRef builds upon the conversational KG-QA benchmark [ConvQuestions](/dataset/convquestions). Questions come from five different domains: books, movies, music, TV serie...
Provide a detailed description of the following dataset: ConvRef
ISAdetect dataset
This repository holds two datasets: one with both the original binaries and the code sections extracted from them (“full dataset”), and one with only the code sections (“only code sections”). The code sections were extracted by carving out sections of the binary that were marked as executable. The binaries were scraped...
Provide a detailed description of the following dataset: ISAdetect dataset
BBBC041
P. vivax (malaria) infected human blood smears with bounding box annotations. The data consists of two classes of uninfected cells (RBCs and leukocytes) and four classes of infected cells (gametocytes, rings, trophozoites, and schizonts).
Provide a detailed description of the following dataset: BBBC041
SoMeSci
Knowledge about software used in scientific investigations is important for several reasons, for instance, to enable an understanding of provenance and methods involved in data handling. However, software is usually not formally cited, but rather mentioned informally within the scholarly description of the investigatio...
Provide a detailed description of the following dataset: SoMeSci
MuDoCo_QueryRewrite
<Task description: joint learning of coreference resolution and query rewrite> Given an ongoing dialogue between a user and a dialogue assistant, for the user query, the model is required to predict both coreference links between the query and the dialogue context, and the self-contained rewritten user query that is...
Provide a detailed description of the following dataset: MuDoCo_QueryRewrite
VerbCL
**VerbCL** is a dataset that consists of the citation graph of court opinions, which cite previously published court opinions in support of their arguments. In particular, it focuses on the verbatim quotes, i.e., where the text of the original opinion is directly reused. **VerbCL** is derived from CourtListener and ...
Provide a detailed description of the following dataset: VerbCL
BugRepo
BugRepo maintains a collection of bug reports that are publicly available for research purposes. Bug reports are a main data source for facilitating NLP-based research in software engineering. We categorize the datasets into the following research directions.
Provide a detailed description of the following dataset: BugRepo
THRED
This is two-hop relation extraction dataset derived from WikiHop dataset [1]. [1] Johannes Welbl and Pontus Stenetorp and Sebastian Riedel. Constructing Datasets for Multi-hop Reading Comprehension Across Documents. TACL, 2018.
Provide a detailed description of the following dataset: THRED
HiXray
**HiXray** is a High-quality X-ray security inspection image dataset, which contains 102,928 common prohibited items of 8 categories. It has been gathered from the real-world airport security inspection and annotated by professional security inspectors
Provide a detailed description of the following dataset: HiXray
Invisible Mobile Keyboard Dataset
**Invisible Mobile Keyboard Dataset** contains user initial, age, type of mobile devices, size of the screen, time taken for typing each phrase, and annotation of typed phrases with coordinate values of the typed position (x and y points). The collected dataset is the first and only dataset for a novel IMK decoding ta...
Provide a detailed description of the following dataset: Invisible Mobile Keyboard Dataset
MSDA
* 5 domains: synthetic domain, document domain, street view domain, handwritten domain, and car license domain * over five million images
Provide a detailed description of the following dataset: MSDA
MAPS
MAPS – standing for MIDI Aligned Piano Sounds – is a database of MIDI-annotated piano recordings. MAPS has been designed in order to be released in the music information retrieval research community, especially for the development and the evaluation of algorithms for single-pitch or multipitch estimation and automatic ...
Provide a detailed description of the following dataset: MAPS
LLVIP
* Visible-infrared Paired Dataset for Low-light Vision * 30976 images (15488 pairs) * 24 dark scenes, 2 daytime scenes * Support for image-to-image translation (visible to infrared, or infrared to visible), visible and infrared image fusion, low-light pedestrian detection, and infrared pedestrian detection * (...
Provide a detailed description of the following dataset: LLVIP
VisEvent
**VisEvent** (Visible-Event benchmark) is a dataset constructed for the evaluation of tracking by combing visible and event cameras. VisEvent is featured in: Large-scale: 820 video sequences (RGB video + Event flows), contains 371,128 frames, 500 / 320 for the train / testing respectively; High-quality Dense An...
Provide a detailed description of the following dataset: VisEvent
FLUE
FLUE is a French Language Understanding Evaluation benchmark. It consists of 5 tasks: Text Classification, Paraphrasing, Natural Language Inference, Constituency Parsing and Part-of-Speech Tagging, and Word Sense Disambiguation.
Provide a detailed description of the following dataset: FLUE
Images from camera traps in the Jura and Ain counties (France)
This dataset contains images taken from camera traps set up in the Jura and Ain counties in France. We use this dataset to illustrate the training of a deep learning algorithm with application to animal specie sidentification. See more here https://github.com/oliviergimenez/computo-deeplearning-occupany-lynx.
Provide a detailed description of the following dataset: Images from camera traps in the Jura and Ain counties (France)
MAST
A new data consolidation called Multi-Attributed and Structured Text-to-face (MAST) dataset. The motivation is to have a large corpus of high-quality face images with fine-grained and attribute-focussed annotations. This has the benefits of the attribute oriented approach as well as the semantics in a textual descripti...
Provide a detailed description of the following dataset: MAST
FlickrStyle10K
FlickrStyle10K is collected and built on Flickr30K image caption dataset. The original FlickrStyle10K dataset has 10,000 pairs of images and stylized captions including humorous and romantic styles. However, only 7,000 pairs from the official training set are now publicly accessible. The dataset can be downloaded via ht...
Provide a detailed description of the following dataset: FlickrStyle10K
PEM Fuel Cell Dataset
This dataset are about Nafion 112 membrane standard tests and MEA activation tests of PEM fuel cell in various operation condition. Dataset include two general electrochemical analysis method, Polarization and Impedance curves. In this dataset, effect of different pressure of H2/O2 gas, different voltages and various h...
Provide a detailed description of the following dataset: PEM Fuel Cell Dataset
KITTI MOTS
The Multi-Object and Segmentation (MOTS) benchmark [2] consists of 21 training sequences and 29 test sequences. It is based on the KITTI Tracking Evaluation 2012 and extends the annotations to the Multi-Object and Segmentation (MOTS) task. To this end, we added dense pixel-wise segmentation labels for every object. We ...
Provide a detailed description of the following dataset: KITTI MOTS
OSLD
Open Set Logo Detection Dataset (OSLD Dataset) is a dataset of eCommerce product images with associated brand logo images. It is released under creative commons (CC BY-NC 4.0) license to promote research in open set logo detection. The dataset can be used only for research purposes. The dataset contains: - 20K eComm...
Provide a detailed description of the following dataset: OSLD
ESPADA
We present a new aerial image dataset, named ESPADA, intended for the training of deep neural networks for depth image estimation from a single aerial image. Given the difficulty of creating aerial image datasets containing image pairs of chromatic images related to their depth images, simulators such as AirSim have be...
Provide a detailed description of the following dataset: ESPADA
AP-10K
AP-10K is the first large-scale benchmark for general animal pose estimation, to facilitate the research in animal pose estimation. AP-10K consists of 10,015 images collected and filtered from 23 animal families and 60 species following the taxonomic rank and high-quality keypoint annotations labeled and checked manual...
Provide a detailed description of the following dataset: AP-10K
N15News
N15News is a large-scale multimodal news dataset comprising 200K imagetext pairs and 15 categories, which exceeding the previous news dataset in both the number of categories and samples. Image source: [https://arxiv.org/pdf/2108.13327v1.pdf](https://arxiv.org/pdf/2108.13327v1.pdf)
Provide a detailed description of the following dataset: N15News
Cats and Dogs
A large set of images of cats and dogs. Homepage: https://www.microsoft.com/en-us/download/details.aspx?id=54765
Provide a detailed description of the following dataset: Cats and Dogs
GESTURES
This is the dataset to support the paper: Fernando Pérez-García et al., 2021, Transfer Learning of Deep Spatiotemporal Networks to Model Arbitrarily Long Videos of Seizures. The paper has been accepted for publication at the 24th International Conference on Medical Image Computing and Computer Assisted Intervention (...
Provide a detailed description of the following dataset: GESTURES
ASR-GLUE
The **ASR-GLUE** benchmark is a collection of 6 different NLU (Natural Language Understanding) tasks for evaluating the performance of models under automatic speech recognition (ASR) error across 3 different levels of background noise and 6 speakers with various voice characteristics.
Provide a detailed description of the following dataset: ASR-GLUE
SHIFT15M
**SHIFT15M** is a dataset that can be used to properly evaluate models in situations where the distribution of data changes between training and testing. The SHIFT15M dataset has several good properties: (i) Multiobjective. Each instance in the dataset has several numerical values that can be used as target variables...
Provide a detailed description of the following dataset: SHIFT15M
VesselGraph
**VesselGraph** is a dataset of whole-brain vessel graphs based on specific imaging protocols. Specifically, vascular graphs are extracted using a refined graph extraction scheme leveraging the volume rendering engine Voreen and provided in an accessible and adaptable form through the OGB and PyTorch Geometric dataload...
Provide a detailed description of the following dataset: VesselGraph
MEDIC
**MEDIC** is a large social media image classification dataset for humanitarian response consisting of 71,198 images to address four different tasks in a multi-task learning setup. It consists data from several data sources such as [CrisisMMD](crisismmd), data from AIDR and Damage Multimodal Dataset (DMD).
Provide a detailed description of the following dataset: MEDIC
CrossedWires
**CrossedWires** is a living dataset of models and hyperparameters that exposes semantic differences between two popular deep learning frameworks: PyTorch and Tensorflow. The CrossedWires dataset currently consists of models trained on CIFAR10 images using three different computer vision architectures: VGG16, ResNe...
Provide a detailed description of the following dataset: CrossedWires
HeadlineCause
**HeadlineCause** is a dataset for detecting implicit causal relations between pairs of news headlines. The dataset includes over 5000 headline pairs from English news and over 9000 headline pairs from Russian news labeled through crowdsourcing. The pairs vary from totally unrelated or belonging to the same general top...
Provide a detailed description of the following dataset: HeadlineCause
TIMo
TIMo (Time-of-Flight Indoor Monitoring) is a dataset of infrared and depth videos intended for the use in Anomaly Detection and Person Detection/People Counting. It features more than 1,500 sequences for anomaly detection, which sum up to more than 500,000 individual frames. For person detection the dataset contains mo...
Provide a detailed description of the following dataset: TIMo
Lyra
Lyra is a dataset for code generation that consists on Python code with embedded SQL. This dataset contains 2,000 carefully annotated database manipulation programs from real usage projects. Each program is paired with both a Chinese comment and an English comment.
Provide a detailed description of the following dataset: Lyra
BSARD
The **Belgian Statutory Article Retrieval Dataset (BSARD)** is a French native corpus for studying *statutory article retrieval*. BSARD consists of more than 22,600 statutory articles from Belgian law and about 1,100 legal questions posed by Belgian citizens and labeled by experienced jurists with relevant articles fro...
Provide a detailed description of the following dataset: BSARD
ReadingBank
ReadingBank is a benchmark dataset for reading order detection built with weak supervision from WORD documents, which contains 500K document images with a wide range of document types as well as the corresponding reading order information.
Provide a detailed description of the following dataset: ReadingBank
WikiNLDB
WikiNLDB is a novel dataset for training Natural Language Databases (NLDBs) which is generated by transforming structured data from Wikidata into natural language facts and queries. Image source: [https://arxiv.org/pdf/2106.01074v1.pdf](https://arxiv.org/pdf/2106.01074v1.pdf)
Provide a detailed description of the following dataset: WikiNLDB
UQ NIDS Datasets
A comprehensive dataset, merging all the aforementioned datasets. The newly published dataset represents the benefits of shared dataset feature sets, where the merging of multiple smaller ones is possible. This will eventually lead to a bigger and more universal NIDS datasets containing flows from multiple network setu...
Provide a detailed description of the following dataset: UQ NIDS Datasets
UQ NetFlow NIDS v1
A comprehensive dataset, merging all the aforementioned datasets. The newly published dataset represents the benefits of shared dataset feature sets, where the merging of multiple smaller ones is possible. This will eventually lead to a bigger and more universal NIDS datasets containing flows from multiple network setu...
Provide a detailed description of the following dataset: UQ NetFlow NIDS v1
AMFDS
# Arabic Multi Fonts Dataset A multi-word multi-font Arabic word-image dataset. AMDS is a dataset of Arabic word images. The dataset was generated using the TextImagesToolkit https://github.com/msfasha/TextImagesToolkit. The database of comprised of a number of binary files and text files. The binary files s...
Provide a detailed description of the following dataset: AMFDS
VideoMatte240K
VideoMatte240K consists of 484 high-resolution green screen videos and generate a total of 240,709 unique frames of alpha mattes and foregrounds with chroma-key software Adobe After Effects. The videos are purchased as stock footage or found as royalty-free materials online. 384 videos are at 4K resolution and 100 are ...
Provide a detailed description of the following dataset: VideoMatte240K
PhotoMatte85
PhotoMatte85 contains 85 protrait images. The dataset is donated to us by a third-party commercial company. The footage are shot with professional studio lighting and the subjects are in standard portrait posing. We provide the alpha matte and foreground images extracted from the green screen photos. Due to license iss...
Provide a detailed description of the following dataset: PhotoMatte85
Phy-Q
**Phy-Q** is a benchmark that requires an agent to reason about physical scenarios and take an action accordingly. Inspired by the physical knowledge acquired in infancy and the capabilities required for robots to operate in real-world environments, the authors identify 15 essential physical scenarios. For each scenari...
Provide a detailed description of the following dataset: Phy-Q
UQ NIDS Datasets (FlowMeter Format)
CICFlowMeter format of the datasets are made up of 83 features.
Provide a detailed description of the following dataset: UQ NIDS Datasets (FlowMeter Format)
CICIDS2018
CICIDS2018 includes seven different attack scenarios: Brute-force, Heartbleed, Botnet, DoS, DDoS, Web attacks, and infiltration of the network from inside. The attacking infrastructure includes 50 machines and the victim organization has 5 departments and includes 420 machines and 30 servers. The dataset includes the c...
Provide a detailed description of the following dataset: CICIDS2018
PCC
The Potsdam Commentary Corpus (PCC) is a corpus of 220 German newspaper commentaries (2.900 sentences, 44.000 tokens) taken from the online issues of the Märkische Allgemeine Zeitung (MAZ subcorpus) and Tagesspiegel (ProCon subcorpus) and is annotated with a range of different types of linguistic information. The ce...
Provide a detailed description of the following dataset: PCC
MASC
The Manually Annotated Sub-Corpus (MASC) consists of approximately 500,000 words of contemporary American English written and spoken data drawn from the Open American National Corpus (OANC). All of MASC includes manually validated annotations for sentence boundaries, token, lemma and POS; noun and verb chunks; name...
Provide a detailed description of the following dataset: MASC
EVIL
To automatically generate Python and assembly programs used for security exploits, we curated a large dataset for feeding NMT techniques. A sample in the dataset consists of a snippet of code from these exploits and their corresponding description in the English language. We collected exploits from publicly available d...
Provide a detailed description of the following dataset: EVIL
FinQA
FinQA is a new large-scale dataset with Question-Answering pairs over Financial reports, written by financial experts. The dataset contains 8,281 financial QA pairs, along with their numerical reasoning processes.
Provide a detailed description of the following dataset: FinQA
Common Objects in 3D
Common Objects in 3D is a large-scale dataset with real multi-view images of object categories annotated with camera poses and ground truth 3D point clouds. The dataset contains a total of 1.5 million frames from nearly 19,000 videos capturing objects from 50 MS-COCO categories and, as such, it is significantly larger ...
Provide a detailed description of the following dataset: Common Objects in 3D
Creative Style Responses
Raw responses of ~10,000 people to a simple survey of creative habits. The numeric responses are an ordinal scale 1-5 for questions that ask about 2 contrasting creative habits/preferences along a given habit dimension. The endpoints of the scale are in the name of the column. The Discipline field is a 'check all that ...
Provide a detailed description of the following dataset: Creative Style Responses
Discipline Mapping
Mapping of detailed discipline tags to one of three broader disciplines (Arts, Science, Business)
Provide a detailed description of the following dataset: Discipline Mapping
Gender Mapping
Mapping of free text gender entries to one of three genders: Male, Female, Non-Binary.
Provide a detailed description of the following dataset: Gender Mapping
Shadow Accrual Maps
Large-scale shadows from buildings in a city play an important role in determining the environmental quality of public spaces. They can be both beneficial, such as for pedestrians during summer, and detrimental, by impacting vegetation and by blocking direct sunlight. Determining the effects of shadows requires the acc...
Provide a detailed description of the following dataset: Shadow Accrual Maps
IfAct
We consider the task of **identifying human actions visible in online videos**. We focus on the widely spread genre of lifestyle vlogs, which consist of videos of people performing actions while verbally describing them. Our goal is to identify if actions mentioned in the speech description of a video are visually pr...
Provide a detailed description of the following dataset: IfAct
Creative Habit Tags
Survey responses where all creative habit ordinal responses were converted to Creative Habit Tags - these tags were used in the analysis to build a network of people linked if they share similar creative habit sets, or a network of creative habits linked if the co-occur in the similar sets of people.
Provide a detailed description of the following dataset: Creative Habit Tags
Business Matching
This is a proprietary dataset from a large internet services company of ranked pairs of relevant and irrelevant businesses for different queries, for a total of 17,069 pairs. How well a query matches a candidate is represented by 41 features.
Provide a detailed description of the following dataset: Business Matching
Wiki Talk Page Comments
This public dataset contains 127,820 comments from Wikipedia Talk Pages labeled with whether or not they are toxic
Provide a detailed description of the following dataset: Wiki Talk Page Comments
W3C Experts
This is a subset of the TREC 2005 enterprise track data, and consists of 48 topics and 200 candidates per topic, with each candidate labeled as an expert or non-expert for the topic. The task is to rank the candidates based on their expertise on a topic, using a corpus of mailing lists from the World Wide Web Consortiu...
Provide a detailed description of the following dataset: W3C Experts
SMAC
The StarCraft Multi-Agent Challenge (SMAC) is a benchmark that provides elements of partial observability, challenging dynamics, and high-dimensional observation spaces. SMAC is built using the StarCraft II game engine, creating a testbed for research in cooperative MARL where each game unit is an independent RL agent.
Provide a detailed description of the following dataset: SMAC
Source Code Tagger Training Set
# Ensemble Tagger Training and Testing Set This data includes two files: The training set used to create the SCANL Ensemble tagger [1] and the "unseen" testing set that includes words from systems that are not available in the training set. These are derived from a prior dataset of [Grammar Patterns](https://github....
Provide a detailed description of the following dataset: Source Code Tagger Training Set
VIVOS
VIVOS is a free Vietnamese speech corpus consisting of 15 hours of recording speech prepared for Automatic Speech Recognition task. The corpus was prepared by AILAB, a computer science lab of VNUHCM - University of Science, with Prof. Vu Hai Quan is the head of. We publish this corpus in hope to attract more scie...
Provide a detailed description of the following dataset: VIVOS
catbAbI QA-mode
We aim to improve the bAbI benchmark as a means of developing intelligent dialogue agents. To this end, we propose concatenated-bAbI (catbAbI): an infinite sequence of bAbI stories. catbAbI is generated from the bAbI dataset and during training, a random sample/story from any task is drawn without replacement and conca...
Provide a detailed description of the following dataset: catbAbI QA-mode
catbAbI LM-mode
We aim to improve the bAbI benchmark as a means of developing intelligent dialogue agents. To this end, we propose concatenated-bAbI (catbAbI): an infinite sequence of bAbI stories. catbAbI is generated from the bAbI dataset and during training, a random sample/story from any task is drawn without replacement and conca...
Provide a detailed description of the following dataset: catbAbI LM-mode
Security of Alerting Authorities in the WWW: Measuring Namespaces, DNSSEC, and Web PKI
This data set includes all raw data (e.g., collected certificates) of the WWW 2021 paper "Security of Alerting Authorities in the WWW: Measuring Namespaces, DNSSEC, and Web PKI". * Current certificates in use by AA hosts. * CT-logged certificates used by AA hosts.
Provide a detailed description of the following dataset: Security of Alerting Authorities in the WWW: Measuring Namespaces, DNSSEC, and Web PKI
The Rise of Certificate Transparency and Its Implications on the Internet Ecosystem
This includes all data from the ACM IMC 2018 paper "The Rise of Certificate Transparency and Its Implications on the Internet Ecosystem".
Provide a detailed description of the following dataset: The Rise of Certificate Transparency and Its Implications on the Internet Ecosystem
SHAD3S
We introduce the SHAD3S dataset, that for a given contour representation of a mesh, under a given illumination condition, provides the illumination masks on the object, a shadow mask on the ground, its diffuse and sketch renders. [Dataset creation code](https://github.com/bvraghav/standible)
Provide a detailed description of the following dataset: SHAD3S
Exposure-Errors
A dataset of over 24,000 images exhibiting the broadest range of exposure values to date with a corresponding properly exposed image.
Provide a detailed description of the following dataset: Exposure-Errors
sRGB2XYZ Dataset
The sRGB2XYZ dataset contains ~1,200 pairs of camera-rendered sRGB and the corresponding scene-referred CIE XYZ images (971 training, 50 validation, and 244 testing images).
Provide a detailed description of the following dataset: sRGB2XYZ Dataset
Raw2raw dataset
This dataset consists of an unpaired and paired set of images captured by two different smartphone cameras: Samsung Galaxy S9 and iPhone X. The unpaired set includes 196 images captured by each smartphone camera (total of 392). The paired set includes 115 pair of images used for testing. In addition to this paired set,...
Provide a detailed description of the following dataset: Raw2raw dataset
Landscape Dataset
Landscape Dataset consists of landscape images collected from Flickr.
Provide a detailed description of the following dataset: Landscape Dataset
Portrait Dataset
A portrait dataset of images collected from Flickr.
Provide a detailed description of the following dataset: Portrait Dataset
DBFC Dataset
This dataset includes Direct Borohydride Fuel Cell (DBFC) impedance and polarization test in anode with Pd/C, Pt/C and Pd decorated Ni–Co/rGO catalysts. In fact, different concentration of Sodium Borohydride (SBH), applied voltages and various anode catalysts loading with explanation of experimental details of electroc...
Provide a detailed description of the following dataset: DBFC Dataset
CREAK
A testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities with commonsense inferences. Image source: [https://arxiv.org/pdf/2109.01653v1.pdf](https://arxiv.org/pdf/2109.01653v1.pdf)
Provide a detailed description of the following dataset: CREAK
CameraFusion
We present a novel approach to reference-based super-resolution (RefSR) with the focus on real-world dual-camera super-resolution (DCSR). This dataset currently consists of 143 pairs of telephoto and wide-angle images in 4K resolution captured by smartphone dual-cameras. See our paper for more details: Dual-Camera Su...
Provide a detailed description of the following dataset: CameraFusion
Story Cloze
Representation and learning of commonsense knowledge is one of the foundational problems in the quest to enable deep language understanding. This issue is particularly challenging for understanding casual and correlational relationships between events. While this topic has received a lot of interest in the NLP communit...
Provide a detailed description of the following dataset: Story Cloze
MSU Shot Boundary Detection Benchmark
This is a dataset for a shot boundary detection task. The dataset contains 2 existing datasets and 19 manually marked up open source videos with a total length of more than 1200 minutes and 10000 scene transitions. The dataset includes different types of videos with different resolutions from 360×288 to 1920×1080 in MP...
Provide a detailed description of the following dataset: MSU Shot Boundary Detection Benchmark
Real-world graphs for betweenness-centrality ranking estimation
The ground truth betweenness-centralities for the real-world graphs are provided by AlGhamdi et al. (2017), which are computed by the parallel implementation of Brandes algorithm on a 96000-core supercomputer. The ground truth scores for the synthetic networks are provided by Fan et al. (2019) and are computed using th...
Provide a detailed description of the following dataset: Real-world graphs for betweenness-centrality ranking estimation
MOD
MOD is a large-scale open-domain multimodal dialogue dataset incorporating abundant Internet memes into utterances. The dataset consists of ∼45K Chinese conversations with ∼606K utterances. Each conversation contains about 13 utterances with about 4 Internet memes on average and each utterance equipped with an Internet...
Provide a detailed description of the following dataset: MOD
EVIL-Encoders
This dataset contains samples to generate Python code for security exploits. In order to make the dataset representative of real exploits, it includes code snippets drawn from exploits from public databases. Differing from general-purpose Python code found in previous datasets, the Python code of real exploits entails ...
Provide a detailed description of the following dataset: EVIL-Encoders
EVIL-Decoders
This is an assembly dataset built on top of Shellcode_IA32, a dataset for automatically generating assembly from natural language descriptions that consists of 3,200 assembly instructions, commented in the English language, which were collected from shellcodes for IA-32 and written for the Netwide Assembler (NASM) for ...
Provide a detailed description of the following dataset: EVIL-Decoders
Failure-Dataset-OpenStack
This failure dataset contains information on the events collected in the OpenStack cloud computing platform during three different campaigns of fault-injection experiments performed with three different workloads.
Provide a detailed description of the following dataset: Failure-Dataset-OpenStack
SemEval-2021 Task 11: NLPContributionGraph
NLPContributionGraph was introduced as Task 11 at SemEval 2021 for the first time. The task is defined on a dataset of Natural Language Processing (NLP) scholarly articles with their contributions structured to be integrable within Knowledge Graph infrastructures such as the Open Research Knowledge Graph. The structure...
Provide a detailed description of the following dataset: SemEval-2021 Task 11: NLPContributionGraph
Chest-Xray8 (COVID-19)
This dataset contains 1125 X-ray images of the studied individuals’ chests, including 125 images labeled as COVID-19, 500 images labeled as pneumonia, and 500 images labeled as no findings.
Provide a detailed description of the following dataset: Chest-Xray8 (COVID-19)
PlantVillage
The PlantVillage dataset consists of 54303 healthy and unhealthy leaf images divided into 38 categories by species and disease.
Provide a detailed description of the following dataset: PlantVillage
S-COCO
Synthetic COCO (S-COCO) is a synthetically created dataset for homography estimation learning. It was introduced by DeTone et al., where the source and target images are generated by duplicating the same COCO image. The source patch $I_S$ is generated by randomly cropping a source candidate at position $p$ with a size ...
Provide a detailed description of the following dataset: S-COCO
PDS-COCO
Photometrically Distorted Synthetic COCO (PDS-COCO) dataset is a synthetically created dataset for homography estimation learning. The idea is exactly the same as in the Synthetic [COCO (S-COCO)](https://paperswithcode.com/dataset/s-coco) dataset with SSD-like image distortion added at the beginning of the whole proced...
Provide a detailed description of the following dataset: PDS-COCO
BioLeaflets
**BioLeaflets** is a biomedical dataset for Data2Text generation. It is a corpus of 1,336 package leaflets of medicines authorised in Europe, which were obtained by scraping the European Medicines Agency (EMA) website. Package leaflets are included in the packaging of medicinal products and contain information to help ...
Provide a detailed description of the following dataset: BioLeaflets
CholecT50
**CholecT50** is a dataset of endoscopic videos of laparoscopic cholecystectomy surgery introduced to enable research on fine-grained action recognition in laparoscopic surgery. It is annotated with triplet information in the form of <instrument, verb, target>. The dataset is a collection of 50 videos consisting of 4...
Provide a detailed description of the following dataset: CholecT50
WTW
**WTW** (Wired Table in the Wild) is a large-scale dataset which includes well-annotated structure parsing of multiple style tables in several scenes like the photo, scanning files, web pages. WTW dataset has 10970 training samples and 3611 testing ones. The test images are divided into 7 challenging categories. ...
Provide a detailed description of the following dataset: WTW
PMPC
**PMPC** (Persona Match on Persona-Chat) is a dataset for Speaker Persona Detection (SPD) which aims to detect speaker personas based on the plain conversational text.
Provide a detailed description of the following dataset: PMPC
MultiEURLEX
**MultiEURLEX** is a multilingual dataset for topic classification of legal documents. The dataset comprises 65k European Union (EU) laws, officially translated in 23 languages, annotated with multiple labels from the EUROVOC taxonomy. The dataset covers 23 official EU languages from 7 language families.
Provide a detailed description of the following dataset: MultiEURLEX
M-PCCD
The emerging MPEG point cloud codecs (V-PCC and G-PCC variants) are assessed, and best practices for rate allocation are investigated [1]. For this purpose, three experiments are conducted. In the first experiment, a rigorous evaluation of the codecs is performed, adopting test conditions dictated by experts of the gro...
Provide a detailed description of the following dataset: M-PCCD
FunKPoint
**FunKPoint** is a dataset for finding correspondences in visual data that has ground truth correspondences for 10 tasks and 20 object categories.
Provide a detailed description of the following dataset: FunKPoint
MFH
The **MFH** dataset is a multi-viewpoint fine-grained hand hygiene dataset. It contains 73,1147 samples in total, which are collected by 6 camera views in 6 different locations. All samples are split into 7 classes in total. MFH dataset is distinguished from existing datasets in three aspects: the large intra-class dif...
Provide a detailed description of the following dataset: MFH
Hummingbird
**Hummingbird** is a dataset to examine stylistic lexical cues from human perception and BERT used to characterize their discrepancy. In HUMMINGBIRD crowd-workers relabeled benchmarking datasets for style classification tasks.
Provide a detailed description of the following dataset: Hummingbird
WhyAct
**WhyAct** is a dataset for identifying human action reasons in online videos, consisting of 1,077 visual actions manually annotated with their reasons.
Provide a detailed description of the following dataset: WhyAct