dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
Frustrated Legislators: Replication data and code
Description: Replication data and code for Aref, S., and Neal, Z.P., "Identifying hidden coalitions in the US House of Representatives by optimally partitioning signed networks based on generalized balance" (2021) Scientific Reports. http://dx.doi.org/10.1038/s41598-021-98139-w
Provide a detailed description of the following dataset: Frustrated Legislators: Replication data and code
TREx-2p
**TREx-2p** is a dataset to probe whether a pretrained LM possesses “indirect” 2-hop knowledge. It is a 2-hop variant of the T-REx dataset. It has been built by manually examining the 2-hop link existing in the knowledge graph of TREx-1p, and select eight 2- hop relation types that make sense to humans
Provide a detailed description of the following dataset: TREx-2p
Pano3D
Pano3D is a new benchmark for depth estimation from spherical panoramas. Its goal is to drive progress for this task in a consistent and holistic manner. The Pano3D 360 depth estimation benchmark provides a standard Matterport3D train and test split, as well as a secondary GibsonV2 partioning for testing and training...
Provide a detailed description of the following dataset: Pano3D
VoicePrivacy 2020
**VoicePrivacy 2020** is a dataset for developing anonymization solutions for speech technology. It is built from subsets of existing datasets such as: [LibriSpeech](librispeech-1), [LibriTTS](libritts), [VoxCeleb1](voxceleb1), [VoxCeleb2](voxceleb1) and [VCTK](vctk).
Provide a detailed description of the following dataset: VoicePrivacy 2020
ImageTBAD
A dataset of A 3D Computed Tomography (CT) image dataset, ImageTBAD, for segmentation of Type-B Aortic Dissection is published. ImageTBAD contains 100 3D Computed Tomography (CT) images, which is of decent size compared with existing medical imaging datasets. ImageTBAD contains a total of 100 3D CTA images gathered ...
Provide a detailed description of the following dataset: ImageTBAD
WebQA
WebQA, is a new benchmark for multimodal multihop reasoning in which systems are presented with the same style of data as humans when searching the web: Snippets and Images. The system must then identify which information is relevant across modalities and combine it with reasoning to answer the query. Systems will be e...
Provide a detailed description of the following dataset: WebQA
MiniF2F
**MiniF2F** is a dataset of formal Olympiad-level mathematics problems statements intended to provide a unified cross-system benchmark for neural theorem proving. The miniF2F benchmark currently targets Metamath, Lean, and Isabelle and consists of 488 problem statements drawn from the AIME, AMC, and the International M...
Provide a detailed description of the following dataset: MiniF2F
mMARCO
**mMARCO** is a multilingual version of the MS MARCO passage ranking dataset comprising 8 languages that was created using machine translation.
Provide a detailed description of the following dataset: mMARCO
AwA Pose
**AwA Pose** is a large scale animal keypoint dataset with ground truth annotations for keypoint detection of quadruped animals from images.
Provide a detailed description of the following dataset: AwA Pose
TREK-150
**TREK-150** is a benchmark dataset for object tracking in First Person Vision (FPV) videos composed of 150 densely annotated video sequences.
Provide a detailed description of the following dataset: TREK-150
OAK
**OAK** is a dataset for online continual object detection benchmark with an egocentric video dataset. OAK adopts the KrishnaCam videos, an ego-centric video stream collected over nine months by a graduate student. OAK provides exhaustive bounding box annotations of 80 video snippets (~17.5 hours) for 105 object catego...
Provide a detailed description of the following dataset: OAK
Depth in the Wild
**Depth in the Wild** is a dataset for single-image depth perception in the wild, i.e., recovering depth from a single image taken in unconstrained settings. It consists of images in the wild annotated with relative depth between pairs of random points.
Provide a detailed description of the following dataset: Depth in the Wild
Fashion-MMT
**Fashion-MNT** is large-scale bilingual product description dataset called Fashion-MMT, which contains over 114k noisy and 40k manually cleaned description translations with multiple product images.
Provide a detailed description of the following dataset: Fashion-MMT
ComSum
ComSum is a data set of 7 million commit messages for text summarization. When documenting commits, software code changes, both a message and its summary are posted. These messages are gathered and filtered to curate developers' work summarization data set.
Provide a detailed description of the following dataset: ComSum
Indiscapes2
Indiscapes2, a new large-scale diverse dataset of Indic manuscripts with semantic layout annotations. Indiscapes2 contains documents from four different historical collections and is 150% larger than its predecessor, Indiscapes.
Provide a detailed description of the following dataset: Indiscapes2
BnB
BnB is a large-scale and diverse in-domain VLN (Vision and Language Navigation) dataset.
Provide a detailed description of the following dataset: BnB
VIL-100
**VIL-100** is a video instance lane detection dataset, which contains 100 videos with in total 10,000 frames, acquired from different real traffic scenarios. All the frames in each video are manually annotated to a high-quality instance-level lane annotation, and a set of frame-level and video-level metrics are inclu...
Provide a detailed description of the following dataset: VIL-100
Automated Evolution of Feature Logging Statement Levels Using Git Histories and Degree of Interest
Logging—used for system events and security breaches to more informational yet essential aspects of software features—is pervasive. Given the high transactionality of today's software, logging effectiveness can be reduced by information overload. Log levels help alleviate this problem by correlating a priority to logs ...
Provide a detailed description of the following dataset: Automated Evolution of Feature Logging Statement Levels Using Git Histories and Degree of Interest
COCO 10% labeled data
Semi-Supervised Object Detection on COCO 10% labeled data
Provide a detailed description of the following dataset: COCO 10% labeled data
VQA-CE
This dataset provides a new split of VQA v2 (similarly to VQA-CP v2), which is built of questions that are hard to answer for biased models. This dataset is designed to penalize biases, and encourage the learning of models that generalize well.
Provide a detailed description of the following dataset: VQA-CE
CodeXGLUE
CodeXGLUE is a benchmark dataset and open challenge for code intelligence. It includes a collection of code intelligence tasks and a platform for model evaluation and comparison. CodeXGLUE stands for General Language Understanding Evaluation benchmark for CODE. It includes 14 datasets for 10 diversified code intelligen...
Provide a detailed description of the following dataset: CodeXGLUE
Medical Wiki Paralell Corpus for Medical Text Simplification
A medical Wiki paralell corpus for medical text simplification.
Provide a detailed description of the following dataset: Medical Wiki Paralell Corpus for Medical Text Simplification
LiDAR-MOS
# Tasks. In moving object segmentation of point cloud sequences, one has to provide motion labels for each point of the test sequences 11-21. Therefore, the input to all evaluated methods is a list of coordinates of the three-dimensional points along with their remission, i.e., the strength of the reflected laser beam...
Provide a detailed description of the following dataset: LiDAR-MOS
Spider-Realistic
Spider-Realistic dataset is used for evaluation in the paper "Structure-Grounded Pretraining for Text-to-SQL". The dataset is created based on the dev split of the Spider dataset (2020-06-07 version from https://yale-lily.github.io/spider). We manually modified the original questions to remove the explicit mention of c...
Provide a detailed description of the following dataset: Spider-Realistic
Infologic sql queries
Sql queries
Provide a detailed description of the following dataset: Infologic sql queries
ProcGen
Procgen Benchmark includes 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning agent learns generalizable skills.
Provide a detailed description of the following dataset: ProcGen
5DOF GB Interpolation
These are larger MATLAB .mat files required for reproducing plots from the sgbaird-5DOF/interp repository for grain boundary property interpolation. gitID-0055bee_uuID-475a2dfd_paper-data6.mat contains multiple trials of five degree-of-freedom interpolation model runs for various interpolation schemes. gpr46883_gitID-b...
Provide a detailed description of the following dataset: 5DOF GB Interpolation
Self-stimulatory Behavior Dataset
Autism Spectrum Disorders (ASD), often referred to as autism, are neurological disorders characterised by deficits in cognitive skills, social and communicative behaviours. A common way of diagnosing ASD is by studying behavioural cues expressed by the children. We introduce a new publicly available dataset (SSBD) o...
Provide a detailed description of the following dataset: Self-stimulatory Behavior Dataset
2017 Robotic Instrument Segmentation Challenge
Segmentation of robotic instruments is an important problem for robotic assisted minimially invasive surgery. It can be used for simple 2D applications such as overlay masking or 2D tracking but also for more complex 3D tasks such as pose estimation. In this challenge we invite applicants to participate in 3 different ...
Provide a detailed description of the following dataset: 2017 Robotic Instrument Segmentation Challenge
Hocalarim: Turkish Student Reviews
We have constructed our dataset by five fields available on the website that were found convenient for the study of student expectations and experience. This includes out-of-five star ratings on easiness, understandability, recitation, accessibility and helpfulness. Average rating was calculated based on these give...
Provide a detailed description of the following dataset: Hocalarim: Turkish Student Reviews
COVID-19 Disinfo
With the emergence of the COVID-19 pandemic, the political and the medical aspects of disinformation merged as the problem got elevated to a whole new level to become the first global infodemic. Fighting this infodemic has been declared one of the most important focus areas of the World Health Organization, with danger...
Provide a detailed description of the following dataset: COVID-19 Disinfo
Waste Classification data
PROBLEM Waste management is a big problem in our country. Most of the wastes end up in landfills. This leads to many issues like: Increase in landfills, Eutrophication, Consumption of toxic waste by animals, Leachate, Increase in toxins, Land, water and air pollution. APPROACH Studied white papers on waste manage...
Provide a detailed description of the following dataset: Waste Classification data
MUC-4
A dataset for evaluate system's understanding of given passages.
Provide a detailed description of the following dataset: MUC-4
BiSECT
**BiSECT** is a dataset for sentence simplification, which is the ability to take a long, complex sentence and split it into shorter sentences, rephrasing as necessary. BiSECT training data consists of 1 million long English sentences paired with shorter, meaning-equivalent English sentences. These were obtained by ext...
Provide a detailed description of the following dataset: BiSECT
HPS Dataset
HPS Dataset is a collection of 3D humans interacting with large 3D scenes (300-1000 $m^2$, up to 2500 $m^2$). The dataset contains images captured from a head-mounted camera coupled with the reference 3D pose and location of the person in a pre-scanned 3D scene. 7 people in 8 large scenes are captured performing acti...
Provide a detailed description of the following dataset: HPS Dataset
ASTE-Data-V2
A benchmark dataset for the Aspect Sentiment Triplet Extraction, an updated version of ASTE-Data-V1.
Provide a detailed description of the following dataset: ASTE-Data-V2
DRKG
Drug Repurposing Knowledge Graph (DRKG) is a comprehensive biological knowledge graph relating genes, compounds, diseases, biological processes, side effects and symptoms. DRKG includes information from six existing databases including DrugBank, Hetionet, GNBR, String, IntAct and DGIdb, and data collected from recent p...
Provide a detailed description of the following dataset: DRKG
CorruptionDataSet
This original data set includes the following four sheets: Sheet 1: Raw Data (the original data set) Sheet 2: Variables (A list with the variables included in the study) Sheet 3: Countries Scientific Relative Production Sheet 4: Correlations
Provide a detailed description of the following dataset: CorruptionDataSet
TRIP
Tiered Reasoning for Intuitive Physics (TRIP) is a novel commonsense reasoning dataset with dense annotations that enable multi-tiered evaluation of machines’ reasoning process. TRIP serves as a benchmark for physical commonsense reasoning that provides traces of reasoning for an end task of plausibility prediction. Th...
Provide a detailed description of the following dataset: TRIP
Graphine
The Graphine dataset contains 2,010,648 terminology definition pairs organized in 227 directed acyclic graphs. Each node in the graph is associated with a terminology and its definition. Terminologies are organized from coarse-grained ones to fine-grained ones in each graph.
Provide a detailed description of the following dataset: Graphine
MuCo-VQA
MuCo-VQA consist of large-scale (3.7M) multilingual and code-mixed VQA datasets in multiple languages: Hindi (hi), Bengali (bn), Spanish (es), German (de), French (fr) and code-mixed language pairs: en-hi, en-bn, en-fr, en-de and en-es. Image source: [https://arxiv.org/pdf/2109.04653v1.pdf](https://arxiv.org/pdf/210...
Provide a detailed description of the following dataset: MuCo-VQA
LIVECell
The **LIVECell (Label-free In Vitro image Examples of Cells)** dataset is a large-scale microscopic image dataset for instance-segmentation of individual cells in 2D cell cultures. LIVECell consists of 5,239 manually annotated, expert-validated, Incucyte HD phase-contrast microscopy images with a total of 1,686,352 ...
Provide a detailed description of the following dataset: LIVECell
Helix
See https://zenodo.org/record/5500215#.YUCgD51Kg2w
Provide a detailed description of the following dataset: Helix
Dataset of 3D Garments with Sewing Patterns
The Dataset contains more than 23500 3D garment models with their corresponding sewing patterns, each representing a unique garment design sampled from one of the 19 different categories. The dataset is suitable for training Deep Learning models to solve a variety of clothing-related tasks.
Provide a detailed description of the following dataset: Dataset of 3D Garments with Sewing Patterns
YorkTag
YorkTag provides pairs of sharp/blurred images containing fiducial markers and is proposed to train and qualitatively and quantitatively evaluate our model.
Provide a detailed description of the following dataset: YorkTag
GMEG-yahoo
Grammatical error correction dataset for text from Yahoo! Answers
Provide a detailed description of the following dataset: GMEG-yahoo
GMEG-wiki
Grammatical error correction dataset for text from Wikipedia.
Provide a detailed description of the following dataset: GMEG-wiki
SituatedQA
**SituatedQA** is an open-retrieval QA dataset where systems must produce the correct answer to a question given the temporal or geographical context. Answers to the same question may change depending on the extralinguistic contexts (when and where the question was asked).
Provide a detailed description of the following dataset: SituatedQA
E-Manual Corpus
**E-Manual Corpus** is a corpus of 307,957 E-manuals, used for pre-training models for Question Answering on e-manuals.
Provide a detailed description of the following dataset: E-Manual Corpus
CelebA-Dialog
The CelebA-Dialog dataset has the following properties: 1) Facial images are annotated with rich fine-grained labels, which classify one attribute into multiple degrees according to its semantic meaning; 2) Accompanied with each image, there are captions describing the attributes and a user request sample. Image sou...
Provide a detailed description of the following dataset: CelebA-Dialog
MLFW
The Masked LFW (MLFW), based on [Cross-Age LFW (CALFW)](https://paperswithcode.com/dataset/calfw) database, is built using a simple but effective tool that generates masked faces from unmasked faces automatically. Image source: [https://arxiv.org/pdf/2109.05804v1.pdf](https://arxiv.org/pdf/2109.05804v1.pdf)
Provide a detailed description of the following dataset: MLFW
VGaokao
**VGaokao** is a verification style reading comprehension dataset designed for native speakers' evaluation.
Provide a detailed description of the following dataset: VGaokao
Implicit Hate
The **Implicit Hate** corpus is a dataset for hate speech detection with fine-grained labels for each message and its implication. This dataset contains 22,056 tweets from the most prominent extremist groups in the United States; 6,346 of these tweets contain implicit hate speech.
Provide a detailed description of the following dataset: Implicit Hate
ZESHEL
ZESHEL is a zero-shot entity linking dataset, which places more emphasis on understanding the unstructured descriptions of entities to resolve the ambiguity of mentions on four unseen domains. This dataset was constructed using Wikias from FANDOM.
Provide a detailed description of the following dataset: ZESHEL
GD-VCR
Geo-Diverse Visual Commonsense Reasoning (GD-VCR) is a new dataset to test vision-and-language models' ability to understand cultural and geo-location-specific commonsense. Image source: [https://arxiv.org/pdf/2109.06860v1.pdf](https://arxiv.org/pdf/2109.06860v1.pdf)
Provide a detailed description of the following dataset: GD-VCR
Harm-C
**Harm-C** is a dataset for detecting harmful memes related to Covid-19.
Provide a detailed description of the following dataset: Harm-C
Commonsense-Dialogues
**Commonsense-Dialogues** is a crowdsourced dataset of ~11K dialogues grounded in social contexts involving utilization of commonsense. The social contexts used were sourced from the train split of the [SocialIQA](social-iqa) dataset, a multiple-choice question-answering based social commonsense reasoning benchmark.
Provide a detailed description of the following dataset: Commonsense-Dialogues
BenchIE
BenchIE: a benchmark and evaluation framework for comprehensive evaluation of OIE systems for English, Chinese and German. In contrast to existing OIE benchmarks, BenchIE takes into account informational equivalence of extractions: our gold standard consists of fact synsets, clusters in which we exhaustively list all s...
Provide a detailed description of the following dataset: BenchIE
DMO
A large scale dataset to pre-train optical flow prediction network. The data are generated from the DAVIS videos using as-rigid-as-possible principle from Deep-matching and MaskRCNN. The dataset has shown better performance compared to the FlyingChairs dataset.
Provide a detailed description of the following dataset: DMO
FlyingChairs
The "Flying Chairs" are a synthetic dataset with optical flow ground truth. It consists of 22872 image pairs and corresponding flow fields. Images show renderings of 3D chair models moving in front of random backgrounds from Flickr. Motions of both the chairs and the background are purely planar.
Provide a detailed description of the following dataset: FlyingChairs
KVQA
It contains manually verified 183K question-answer pairs about more than 18K persons and 24K images. The questions in this dataset require multi-entity, multi-relation and multi-hop reasoning over KG to arrive at an answer. To enable visual named entity linking, it also provides a support set containing reference image...
Provide a detailed description of the following dataset: KVQA
WADS
Collected in the snow belt region of Michigan's Upper Peninsula, WADS is the first multi-modal dataset featuring dense point-wise labeled sequential LiDAR scans collected in severe winter weather. Over 26 TB of multi modal data has been collected of which over 7 GB of LiDAR point clouds (3.6 billion points) have bee...
Provide a detailed description of the following dataset: WADS
Nelson-Plosser
US Macroeconomic dataset containing 14 time series of monthly observations. They have various lengths but all end in 1988. The variables: consumer price index, industrial production, nominal GNP, velocity, employment, interest rate, nominal wages, GNP deflator, money stock, real GNP, stock prices (S&P500), GNP per capi...
Provide a detailed description of the following dataset: Nelson-Plosser
BioLAMA
**BioLAMA** is a benchmark comprised of 49K biomedical factual knowledge triples for probing biomedical Language Models. It is used to assess the capabilities of Language Models for being valid biomedical knowledge bases.
Provide a detailed description of the following dataset: BioLAMA
BLANCA
**BLANCA** (Benchmarks for LANguage models on Coding Artifacts) is a collection of benchmarks that assess code understanding based on tasks such as predicting the best answer to a question in a forum post, finding related forum posts, or predicting classes related in a hierarchy from class documentation.
Provide a detailed description of the following dataset: BLANCA
ELITR ECA
The ELITR ECA corpus is a multilingual corpus derived from publications of the European Court of Auditors. We use automatic translation together with Bleualign to identify parallel sentence pairs in all 506 translation directions. The result is a corpus comprising 264k document pairs and 41.9M sentence pairs. Descri...
Provide a detailed description of the following dataset: ELITR ECA
MindCraft
**MindCraft** is a fine-grained dataset of collaborative tasks performed by pairs of human subjects in the 3D virtual blocks world of Minecraft. It provides information that captures partners' beliefs of the world and of each other as an interaction unfolds, bringing abundant opportunities to study human collaborative ...
Provide a detailed description of the following dataset: MindCraft
M5Product
The **M5Product** dataset is a large-scale multi-modal pre-training dataset with coarse and fine-grained annotations for E-products. • 6 Million multi-modal samples, 5k properties with 24 Million values • 5 modalities-image text table video audio • 6 Million category annotations with 6k classes • Wide data ...
Provide a detailed description of the following dataset: M5Product
Roof-Image Dataset
We created a building-image paired dataset that contains more than 3K samples using our roof modeling tools. Image source: [https://github.com/llorz/SGA21_roofOptimization/tree/main/RoofGraphDataset](https://github.com/llorz/SGA21_roofOptimization/tree/main/RoofGraphDataset)
Provide a detailed description of the following dataset: Roof-Image Dataset
AnlamVer
In this paper, we present AnlamVer, which is a semantic model evaluation dataset for Turkish designed to evaluate word similarity and word relatedness tasks while discriminating those two relations from each other. Our dataset consists of 500 word-pairs annotated by 12 human subjects, and each pair has two distinct sco...
Provide a detailed description of the following dataset: AnlamVer
EDGAR10-Q Dataset
This dataset is built from 10-Q documents (Quarterly Reports) of publicly listed companies on the SEC.
Provide a detailed description of the following dataset: EDGAR10-Q Dataset
ChFinAnn
Ten years (2008-2018) ChFinAnn documents and human-summarized event knowledge bases to conduct the DS-based event labeling. Five event types included: Equity Freeze (EF), Equity Repurchase (ER), Equity Underweight (EU), Equity Overweight (EO) and Equity Pledge (EP), which belong to major events required to be disclose...
Provide a detailed description of the following dataset: ChFinAnn
TruthfulQA
TruthfulQA is a benchmark to measure whether a language model is truthful in generating answers to questions. The benchmark comprises 817 questions that span 38 categories, including health, law, finance and politics. The authors crafted questions that some humans would answer falsely due to a false belief or misconcep...
Provide a detailed description of the following dataset: TruthfulQA
SWDE
This dataset is a real-world web page collection used for research on the automatic extraction of structured data (e.g., attribute-value pairs of entities) from the Web. We hope it could serve as a useful benchmark for evaluating and comparing different methods for structured web data extraction.
Provide a detailed description of the following dataset: SWDE
wikiHow-image
The dataset consists of 53,189 wikiHow articles across various categories of everyday tasks, 155,265 methods, and 772,294 steps with corresponding images.
Provide a detailed description of the following dataset: wikiHow-image
UESTC RGB-D
UESTC RGB-D Varying-view action database contains 40 categories of aerobic exercise. We utilized 2 Kinect V2 cameras in 8 fixed directions and 1 round direction to capture these actions with the data modalities of RGB video, 3D skeleton sequences and depth map sequences.
Provide a detailed description of the following dataset: UESTC RGB-D
NLB
**Neural Latents** is a benchmark for latent variable modeling of neural population activity. It consists of four datasets of neural spiking activity from cognitive, sensory, and motor areas to promote models that apply to the wide variety of activity seen across these areas.
Provide a detailed description of the following dataset: NLB
HM3D
**Habitat-Matterport 3D** (HM3D) is a large-scale dataset of 1,000 building-scale 3D reconstructions from a diverse set of real-world locations. Each scene in the dataset consists of a textured 3D mesh reconstruction of interiors such as multi-floor residences, stores, and other private indoor spaces. HM3D surpasses...
Provide a detailed description of the following dataset: HM3D
Depth VIDIT
VIDIT is a reference evaluation benchmark and to push forward the development of illumination manipulation methods. Virtual datasets are not only an important step towards achieving real-image performance but have also proven capable of improving training even when real datasets are possible to acquire and available. V...
Provide a detailed description of the following dataset: Depth VIDIT
ADEFAN
This data set contains 50 low resolution (640 x 360) short videos containing a variety real life activities.
Provide a detailed description of the following dataset: ADEFAN
VR Curve on Surface Drawing Dataset
The datasets includes curves drawn on 3D surfaces (triangle meshes) in Virtual Reality. A total of 2,880 curves were created using two different techniques by 20 users on 6 meshes. For each curve, a 3D curve executed by the user is provided, the projected curve created on the mesh, and the ground truth target curve on ...
Provide a detailed description of the following dataset: VR Curve on Surface Drawing Dataset
Machine Learning Quantum Reaction Rate Constants
Dataset of 1,517,419 quantum reaction rate constant products kQM(T)QR(T) computed from the transmission coefficient for model single and double barrier minimum energy paths. Here kQM(T) is the quantum reaction rate constant at temperature T and QR(T) is the reactant partition function computed with the rigid rotor and ...
Provide a detailed description of the following dataset: Machine Learning Quantum Reaction Rate Constants
ReaSCAN
ReaSCAN is a synthetic navigation task that requires models to reason about surroundings over syntactically difficult languages.
Provide a detailed description of the following dataset: ReaSCAN
CMU Motion Capture
This dataset of motions is free for all uses. Please don't crawl this database! Check out the FAQs. This data is free for use in research projects. You may include this data in commercially-sold products, but you may not resell this data directly, even in converted form. If you publish results obtained using t...
Provide a detailed description of the following dataset: CMU Motion Capture
Berkeley MHAD
Description The Berkeley Multimodal Human Action Database (MHAD) contains 11 actions performed by 7 male and 5 female subjects in the range 23-30 years of age except for one elderly subject. All the subjects performed 5 repetitions of each action, yielding about 660 action sequences which correspond to about 82 minute...
Provide a detailed description of the following dataset: Berkeley MHAD
Novel COVID-19 Chestxray Repository
##_Authors of the Dataset_: - Pratik Bhowal (B.E., Dept of Electronics and Instrumentation Engineering, Jadavpur University Kolkata, India) [[LinkedIn]](https://www.linkedin.com/in/pratik-bhowal-1066aa198?lipi=urn%3Ali%3Apage%3Ad_flagship3_profile_view_base_contact_details%3B%2BqgwqwxJRIep5K454MTQ6w%3D%3D), [[Github]]...
Provide a detailed description of the following dataset: Novel COVID-19 Chestxray Repository
EmoCause
**EmoCause** is a dataset of annotated emotion cause words in emotional situations from the [EmpatheticDialogues](/dataset/empatheticdialogues) valid and test set. The goal is to recognize emotion cause words in sentences by training only on sentence-level emotion labels without word-level labels (i.e., weakly-supervis...
Provide a detailed description of the following dataset: EmoCause
CodeQA
CodeQA is a free-form question answering dataset for the purpose of source code comprehension: given a code snippet and a question, a textual answer is required to be generated. CodeQA contains a Java dataset with 119,778 question-answer pairs and a Python dataset with 70,085 question-answer pairs. Description from...
Provide a detailed description of the following dataset: CodeQA
Draper VDisc Dataset
Draper VDISC Dataset - Vulnerability Detection in Source Code The dataset consists of the source code of 1.27 million functions mined from open source software, labeled by static analysis for potential vulnerabilities. For more details on the dataset and benchmark results, see https://arxiv.org/abs/1807.04320. Th...
Provide a detailed description of the following dataset: Draper VDisc Dataset
VISUELLE
VISUELLE is a repository build upon the data of a real fast fashion company, Nunalie, and is composed of 5577 new products and about 45M sales related to fashion seasons from 2016-2019. Each product in VISUELLE is equipped with multimodal information: its image, textual metadata, sales after the first release date, and...
Provide a detailed description of the following dataset: VISUELLE
ARCA23K
ARCA23K is a dataset of labelled sound events created to investigate real-world label noise. It contains 23,727 audio clips originating from Freesound, and each clip belongs to one of 70 classes taken from the AudioSet ontology. The dataset was created using an entirely automated process with no manual verification of ...
Provide a detailed description of the following dataset: ARCA23K
EntityQuestions
**EntityQuestions** is a dataset of simple, entity-rich questions based on facts from Wikidata (e.g., "Where was Arve Furset born? ").
Provide a detailed description of the following dataset: EntityQuestions
OPV2V
**OPV2V** is a large-scale open simulated dataset for Vehicle-to-Vehicle perception. It contains over 70 interesting scenes, 11,464 frames, and 232,913 annotated 3D vehicle bounding boxes, collected from 8 towns in CARLA and a digital town of Culver City, Los Angeles.
Provide a detailed description of the following dataset: OPV2V
ObjectFolder
**ObjectFolder** is a dataset for multisensory object-centric perception, reasoning, and interaction. It consists of 100 virtualized objects. ObjectFolder encodes the visual, auditory, and tactile sensory data for all objects, enabling a number of multisensory object recognition tasks.
Provide a detailed description of the following dataset: ObjectFolder
Bentham
Bentham manuscripts refers to a large set of documents that were written by the renowned English philosopher and reformer Jeremy Bentham (1748-1832). Volunteers of the Transcribe Bentham initiative transcribed this collection. Currently, >6 000 documents or > 25 000 pages have been transcribed using this public web pla...
Provide a detailed description of the following dataset: Bentham
Saint Gall
Saint Gall dataset contains handwritten historical manuscripts written in Latin that date back to the 9th century. It consists of 60 pages, 1 410 text lines and 11 597 words.
Provide a detailed description of the following dataset: Saint Gall
Konzil
Konzil dataset was created by specialists of the University of Greifswald. It contains manuscripts written in modern German. Train sample consists of 353 lines, validation - 29 lines and test - 87 lines.
Provide a detailed description of the following dataset: Konzil
Schiller
Schiller contains handwritten texts written in modern German. Train sample consists of 244 lines, validation - 21 lines and test - 63 lines.
Provide a detailed description of the following dataset: Schiller
Ricordi
Ricordi contains handwritten texts written in Italian. Train sample consists of 295 lines, validation - 19 lines and test - 69 lines.
Provide a detailed description of the following dataset: Ricordi
Patzig
Patzig contains handwritten texts written in modern German. Train sample consists of 485 lines, validation - 38 lines and test -118 lines.
Provide a detailed description of the following dataset: Patzig