dataset_name
stringlengths
2
128
description
stringlengths
1
9.7k
prompt
stringlengths
59
185
TCC
The largest and most realistic dataset available for TCC. It consists of 600 real-world videos recorded with a high-resolution mobile phone camera shooting 1824 x 1368 sized pictures. The length of these videos ranges from 3 to 17 frames (7.3 on average, the median is 7.0 and mode is 8.5). Ground truth information is p...
Provide a detailed description of the following dataset: TCC
USPTO-50k
Subset and preprocessed version of Chemical reactions from US patents (1976-Sep2016) by Daniel Lowe. It includes 50K randomly selected reactions that was later classified into 10 reaction classes by Nadine Schneider et al.
Provide a detailed description of the following dataset: USPTO-50k
Eduge
Eduge news classification dataset provided by Bolorsoft LLC. Used to train the Eduge.mn production news classifier 75K news articles in 9 categories: урлаг соёл, эдийн засаг, эрүүл мэнд, хууль, улс төр, спорт, технологи, боловсрол and байгал орчин Download train/test data via: [train](https://storage.googleapis....
Provide a detailed description of the following dataset: Eduge
Learning to Autofocus
This dataset contains 510 focal stacks (49 different focal distances) from in-the-wild scenes with calculated depth from SFM. This dataset was designed for research on Autofocus but can be used for any research which is interested in focal stacks, defocus cues, or depth signals (particularly for interest in close depth...
Provide a detailed description of the following dataset: Learning to Autofocus
CMeEE
Chinese Medical Named Entity Recognition, a dataset first released in CHIP20204, is used for CMeEE task. Given a pre-defined schema, the task is to identify and extract entities from the given sentence and classify them into nine categories: disease, clinical manifestations, drugs, medical equipment, medical procedures...
Provide a detailed description of the following dataset: CMeEE
CMeIE
Chinese Medical Information Extraction, a dataset that is also released in CHIP2020, is used for CMeIE task. The task is aimed at identifying both entities and relations in a sentence following the schema constraints. There are 53 relations defined in the dataset, including 10 synonymous sub-relationships and 43 other ...
Provide a detailed description of the following dataset: CMeIE
CHIP-STS
CHIP Semantic Textual Similarity, a dataset for sentence similarity in the non-i.i.d. (non-independent and identically distributed) setting, is used for the CHIP-STS task. Specifically, the task aims to transfer learning between disease types on Chinese disease questions and answer data. Given question pairs related...
Provide a detailed description of the following dataset: CHIP-STS
CHIP-CDN
CHIP Clinical Diagnosis Normalization, a dataset that aims to standardize the terms from the final diagnoses of Chinese electronic medical records, is used for the CHIP-CDN task. Given the original phrase, the task is required to normalize it to standard terminology based on the International Classification of Disea...
Provide a detailed description of the following dataset: CHIP-CDN
CHIP-CTC
CHIP Clinical Trial Classification, a dataset aimed at classifying clinical trials eligibility criteria, which are fundamental guidelines of clinical trials defined to identify whether a subject meets a clinical trial or not, is used for the CHIP-CTC task. All text data are collected from the website of the Chinese C...
Provide a detailed description of the following dataset: CHIP-CTC
KUAKE-QIC
KUAKE Query Intent Classification, a dataset for intent classification, is used for the KUAKE-QIC task. Given the queries of search engines, the task requires to classify each of them into one of 11 medical intent categories defined in KUAKE-QIC, including diagnosis, etiology analysis, treatment plan, medical advice...
Provide a detailed description of the following dataset: KUAKE-QIC
KUAKE-QTR
KUAKE Query Title Relevance, a dataset used to estimate the relevance of the title of a query document, is used for the KUAKE-QTR task. Given a query (e.g., “Symptoms of vitamin B deficiency”), the task aims to find the relevant title (e.g., “The main manifestations of vitamin B deficiency”).
Provide a detailed description of the following dataset: KUAKE-QTR
KUAKE-QQR
KUAKE Query-Query Relevance, a dataset used to evaluate the relevance of the content expressed in two queries, is used for the KUAKE-QQR task. Similar to KUAKE-QTR, the task aims to estimate query-query relevance, which is an essential and challenging task in real-world search engines.
Provide a detailed description of the following dataset: KUAKE-QQR
Oxford Road Boundaries
The **Oxford Road Boundaries** is a dataset designed for training and testing machine-learning-based road-boundary detection and inference approaches. The authors have hand-annotated two of the 10 km-long forays from the [Oxford Robotcar Dataset](oxford-robotcar-dataset) and generated from other forays several thou...
Provide a detailed description of the following dataset: Oxford Road Boundaries
JRDB-Act
**JRDB-Act** is an extension of the [JRDB](jrdb) dataset to create a large-scale multi-modal dataset for spatio-temporal action, social group and activity detection. JRDB-Act has been densely annotated with atomic actions, comprises over 2.8M action labels, constituting a large-scale spatio-temporal action detectio...
Provide a detailed description of the following dataset: JRDB-Act
RyanSpeech
**RyanSpeech** is a speech corpus for research on automated text-to-speech (TTS) systems. This dataset contains textual materials from real-world conversational settings. These materials contain over 10 hours of a professional male voice actor's speech recorded at 44.1 kHz.
Provide a detailed description of the following dataset: RyanSpeech
LARC
**LARC** is a dataset built from ARC (Abstraction and Reasoning Corpus). ARC is a set of tasks that tests an agent's ability to flexibly solve novel problems. While most ARC tasks are easy for humans, they are challenging for state-of-the-art AI. LARC or Language-annotated ARC, is a collection of natural language de...
Provide a detailed description of the following dataset: LARC
Physion
**Physion** is a visual and physical prediction benchmark to measure the performance of machine learning models on making predictions about commonplace real world physical events. In realistically simulating a wide variety of physical phenomena -- rigid and soft-body collisions, stable multi-object configurations, roll...
Provide a detailed description of the following dataset: Physion
EuroCrops
EuroCrops is a dataset for automatic vegetation classification from multi-spectral and multi-temporal satellite data, annotated with official LIPS reporting data from countries of the European Union, curated by the Technical University of Munich and GAF AG. The project is managed by the DLR Space Administration and fun...
Provide a detailed description of the following dataset: EuroCrops
GigaSpeech
GigaSpeech, an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised and unsupervised training.
Provide a detailed description of the following dataset: GigaSpeech
Imgur5K
**Imgur5k** is a large-scale handwritten in-the-wild dataset, containing challenging real world handwritten samples from nearly 5K writers. It consists of ~135K handwritten English words from 5K different images. As opposed to existing dataests for OCR which have limited variability in their images, the images in Imgur...
Provide a detailed description of the following dataset: Imgur5K
DisKnE
**DisKnE** is a benchmark for Disease Knowledge Evaluation built from MedNLI and MEDIQA-NLI. This benchmark is constructed to specifically test the medical reasoning capabilities of ML models, such as mapping symptoms to diseases. The dataset was built by annotating each positive MedNLI example with the types of med...
Provide a detailed description of the following dataset: DisKnE
PATTERN
PATTERN is a node classification tasks generated with [Stochastic Block Models](https://paperswithcode.com/paper/community-detection-and-stochastic-block), which is widely used to model communities in social networks by modulating the intra- and extra-communities connections, thereby controlling the difficulty of the t...
Provide a detailed description of the following dataset: PATTERN
CLUSTER
CLUSTER is a node classification tasks generated with [Stochastic Block Models](https://paperswithcode.com/paper/community-detection-and-stochastic-block), which is widely used to model communities in social networks by modulating the intra- and extra-communities connections, thereby controlling the difficulty of the t...
Provide a detailed description of the following dataset: CLUSTER
CSL
CSL is a synthetic dataset introduced in [Murphy et al. (2019)](https://paperswithcode.com/paper/relational-pooling-for-graph-representations) to test the expressivity of GNNs. In particular, graphs are isomorphic if they have the same degree and the task is to classify non-isomorphic graphs.
Provide a detailed description of the following dataset: CSL
Large-scale Anomaly Detection
**Large-scale Anomaly Detection** (LAD) is a database to benchmark anomaly detection in video sequences, which is featured in two aspects. 1) It contains 2000 video sequences including normal and abnormal video clips with 14 anomaly categories including crash, fire, violence, etc. with large scene varieties, making it ...
Provide a detailed description of the following dataset: Large-scale Anomaly Detection
Counting Probe
Probing cross-modal capabilities of Vision & Language models with a counting task. * binary classification * following a FOIL setup (as introduced by Shekhar et al. 2017: https://www.aclweb.org/anthology/P17-1024/)
Provide a detailed description of the following dataset: Counting Probe
BestRev
Survey instrument, analysis code, and anonymized responses for the paper on review practices in SE.
Provide a detailed description of the following dataset: BestRev
COVID-19 Case Surveillance Public Use Data
This case surveillance public use dataset has 12 elements for all COVID-19 cases shared with CDC and includes demographics, any exposure history, disease severity indicators and outcomes, presence of any underlying medical conditions and risk behaviors, and no geographic data.
Provide a detailed description of the following dataset: COVID-19 Case Surveillance Public Use Data
BABEL
BABEL is a large dataset with language labels describing the actions being performed in mocap sequences. BABEL consists of action labels for about 43 hours of mocap sequences from AMASS. Action labels are at two levels of abstraction -- sequence labels describe the overall action in the sequence, and frame labels desc...
Provide a detailed description of the following dataset: BABEL
IFCNet
The full **IFCNet** dataset currently consists of 19,000 CAD models distributed over 65 classes according to the taxonomy of the Industry Foundation Classes (IFC) standard. The IFC standard provides an open data exchange format for projects in the Architecture, Engineering and Construction (AEC) domain. Due to high imb...
Provide a detailed description of the following dataset: IFCNet
FIN
A dataset of financial agreements made public through U.S. Security and Exchange Commission (SEC) filings. Eight documents (totalling 54,256 words) were randomly selected for manual annotation, based on the four NE types provided in the CoNLL-2003 dataset: LOCATION (LOC), ORGANISATION (ORG), PERSON (PER), and MISCELLAN...
Provide a detailed description of the following dataset: FIN
WebVid
WebVid contains 10 million video clips with captions, sourced from the web. The videos are diverse and rich in their content. Both the full 10M set and a 2.5M subset is available for download: https://github.com/m-bain/webvid-dataset
Provide a detailed description of the following dataset: WebVid
Kinships
The Kinships dataset describes relationships between members of the Australian tribe Alyawarra and consists of 10,686 triples. It contains 104 entities representing members of the tribe and 26 relationship types that represent kinship terms such as Adiadya or Umbaidya.
Provide a detailed description of the following dataset: Kinships
CI-MNIST
**CI-MNIST** (Correlated and Imbalanced MNIST) is a variant of [MNIST](/dataset/mnist) dataset with introduced different types of correlations between attributes, dataset features, and an artificial eligibility criterion. For an input image $x$, the label $y \in \\{1, 0\\}$ indicates eligibility or ineligibility, respe...
Provide a detailed description of the following dataset: CI-MNIST
Box2D
Continuous control tasks in the Box2D simulator.
Provide a detailed description of the following dataset: Box2D
JSRT
The standard digital image database with and without chest lung nodules (JSRT database) was created(*1) by the Japanese Society of Radiological Technology (JSRT) in cooperation with the Japanese Radiological Society (JRS) in 1998. Since then, the JSRT database has been used by a number of researchers in the world for v...
Provide a detailed description of the following dataset: JSRT
Sports10
- Games dataset containing 100,000 Gameplay Images of 175 Video Games across 10 Sports Genres - AMERICAN FOOTBALL, BASKETBALL, BIKE RACING, CAR RACING, FIGHTING, HOCKEY, SOCCER, TABLE TENNIS, TENNIS. - Hand-curated images to remove menu/transition frames and only include gameplay sequences. - Games are divided i...
Provide a detailed description of the following dataset: Sports10
Solar-Power
Solar Power Data for Integration Studies NREL's Solar Power Data for Integration Studies are synthetic solar photovoltaic (PV) power plant data points for the United States representing the year 2006. The data are intended for use by energy professionals—such as transmission planners, utility planners, project deve...
Provide a detailed description of the following dataset: Solar-Power
ZhihuRec
ZhihuRec dataset is collected from a knowledge-sharing platform (Zhihu), which is composed of around 100M interactions collected within 10 days, 798K users, 165K questions, 554K answers, 240K authors, 70K topics, and more than 501K user query logs. There are also descriptions of users, answers, questions, authors, and ...
Provide a detailed description of the following dataset: ZhihuRec
Personal Events in Dialogue Corpus
The PEDC is a corpus of 14 episodes of This American Life podcast transcripts that have been annotated for events. The corpus contains excerpts from these episodes (listed in Tabe 1) that are dialogue. The granularity of annotation in this corpus is the token; each token is either annotated as an event, or a nonevent. ...
Provide a detailed description of the following dataset: Personal Events in Dialogue Corpus
MeshRIR
MeshRIR is a dataset of acoustic room impulse responses (RIRs) at finely meshed grid points. Two subdatasets are currently available: one consists of IRs in a 3D cuboidal region from a single source, and the other consists of IRs in a 2D square region from an array of 32 sources. This dataset is suitable for evaluating...
Provide a detailed description of the following dataset: MeshRIR
ONCE
ONCE (One millioN sCenEs) is a dataset for 3D object detection in the autonomous driving scenario. The ONCE dataset consists of 1 million LiDAR scenes and 7 million corresponding camera images. The data is selected from 144 driving hours, which is 20x longer than other 3D autonomous driving datasets available like [nuS...
Provide a detailed description of the following dataset: ONCE
SODA10M
**SODA10M** is a large-scale object detection benchmark for standardizing the evaluation of different self-supervised and semi-supervised approaches by learning from raw data. SODA10M contains 10 million unlabeled images and 20K images labeled with 6 representative object categories. To improve diversity, the images ar...
Provide a detailed description of the following dataset: SODA10M
Calliar
Calliar is a dataset for Arabic calligraphy. The dataset consists of 2500 json files that contain strokes manually annotated for Arabic calligraphy.
Provide a detailed description of the following dataset: Calliar
HICRD
**HICRD** (Heron Island Coral Reef Dataset) is a large-scale real underwater image dataset for underwater image restoration. There are 2000 reference restored images and 6003 original underwater images in the unpaired training set.
Provide a detailed description of the following dataset: HICRD
GOLOS
**Golos** is a Russian speech dataset suitable for speech research. The dataset mainly consists of recorded audio files manually annotated on the crowd-sourcing platform. The total duration of the audio is about 1240 hours. ## **Dataset structure** | Domain | Train files | Train hours | Test files | Te...
Provide a detailed description of the following dataset: GOLOS
JerichoWorld
**JerichoWorld** is a dataset that enables the creation of learning agents that can build knowledge graph-based world models of interactive narratives. Interactive narratives -- or text-adventure games -- are partially observable environments structured as long puzzles or quests in which an agent perceives and interac...
Provide a detailed description of the following dataset: JerichoWorld
X-Fact
X-FACT is a large publicly available multilingual dataset for factual verification of naturally existing real-world claims. The dataset contains short statements in 25 languages and is labeled for veracity by expert fact-checkers. The dataset includes a multilingual evaluation benchmark that measures both out-of-domain...
Provide a detailed description of the following dataset: X-Fact
DocNLI
**DocNLI** is a large-scale dataset for document-level NLI. DocNLI is transformed from a broad range of NLP problems and covers multiple genres of text. The premises always stay in the document granularity, whereas the hypotheses vary in length from single sentences to passages with hundreds of words. Additionally, Doc...
Provide a detailed description of the following dataset: DocNLI
EMOVIE
**EMOVIE** is a Mandarin emotion speech dataset including 9,724 samples with audio files and its emotion human-labeled annotation.
Provide a detailed description of the following dataset: EMOVIE
DISC21
**DISC21** is a benchmark for large-scale image similarity detection. This benchmark is used for the Image Similarity Challenge at NeurIPS'21 (ISC2021). The goal is to determine whether a query image is a modified copy of any image in a reference corpus of size 1~million. The benchmark features a variety of image trans...
Provide a detailed description of the following dataset: DISC21
Hi-Phy
**Hi-Phy** is a benchmark for physical reasoning that allows researchers to test individual physical reasoning capabilities. Inspired by how humans acquire these capabilities, the benchmark proposes a general hierarchy of physical reasoning capabilities with increasing complexity. this benchmark tests capabilities acco...
Provide a detailed description of the following dataset: Hi-Phy
VAW
**VAW** is a large scale visual attributes dataset with explicitly labelled positive and negative attributes. Details: * 620 Unique Attributes including color, shape, texture, posture and many others * 2,260 Unique Objects observed in the wild * 72,274 Images from the Visual Genome Dataset * 4 different eval...
Provide a detailed description of the following dataset: VAW
IMFW
Indian Masked faces in the wild Database is collected into three sets:(i) Indian Celebrity, (ii) Instagram and (iii) Indian Crowd. The Indian Celebrity contains 40 Indian celebrities with 435 images, including Bollywood actors/actresses, television stars, sports personalities, and politicians. The Instagram set contain...
Provide a detailed description of the following dataset: IMFW
Fishnet Open Images
**Fishnet Open Images Database** is a large dataset of EM imagery for fish detection and fine-grained categorisation onboard commercial fishing vessels. The dataset consists of 86,029 images containing 34 object classes, making it the largest and most diverse public dataset of fisheries EM imagery to-date. It includes ...
Provide a detailed description of the following dataset: Fishnet Open Images
Synthetic COVID-19 Chest X-ray
The **Synthetic COVID-19 Chest X-ray Dataset** consists of 21,295 synthetic COVID-19 chest X-ray images to be used for computer-aided diagnosis. These images, generated via an unsupervised domain adaptation approach, are of high quality.
Provide a detailed description of the following dataset: Synthetic COVID-19 Chest X-ray
Dataset for methane combustion
The dataset contains 578,731 structures for methane combustion and their energies and forces under MN15/6-31G** level.
Provide a detailed description of the following dataset: Dataset for methane combustion
CICIDS2017
Intrusion Detection Evaluation Dataset (CIC-IDS2017) Intrusion Detection Systems (IDSs) and Intrusion Prevention Systems (IPSs) are the most important defense tools against the sophisticated and ever-growing network attacks. Due to the lack of reliable test and validation datasets, anomaly-based intrusion detection ap...
Provide a detailed description of the following dataset: CICIDS2017
zbMATH Open dataset 2021
zbMATH Open contains over 4 million bibliographic entries with reviews or abstracts drawn from more than 3.000 journals and book series and more than 190.000 books.
Provide a detailed description of the following dataset: zbMATH Open dataset 2021
Photozilla
**Photozilla** is a large-scale dataset which includes over 990k images belonging to 10 different photographic styles. The dataset can be used to train classification models to automatically classify the images into the relevant style.
Provide a detailed description of the following dataset: Photozilla
Text-to-3D House Model
The dataset contains 2,000 houses, 13,478 rooms and 873 (some rooms have same textures so this number is smaller than the total number of rooms.) texture images with corresponding natural language descriptions. These descriptions are firstly generated from some pre-defined templates and then refined by human workers. T...
Provide a detailed description of the following dataset: Text-to-3D House Model
EchoCP
**EchoCP** is an echocardiography dataset in cTTE targeting PFO (Patent foramen ovale) diagnosis. EchoCP consists of 30 patients with both rest and Valsalva maneuver videos which covers various PFO grades. Patent foramen ovale (PFO) is a potential separation between the septum, primum and septum secundum located in ...
Provide a detailed description of the following dataset: EchoCP
WikiPII
WikiPII, an automatically labeled dataset composed of Wikipedia biography pages, annotated for personal information extraction.
Provide a detailed description of the following dataset: WikiPII
NVGaze
Quality, diversity, and size of training dataset are critical factors for learning-based gaze estimators. We create two datasets satisfying these criteria for near-eye gaze estimation under infrared illumination: a synthetic dataset using anatomically-informed eye and face models with variations in face shape, gaze dir...
Provide a detailed description of the following dataset: NVGaze
UA-GEC
UA-GEC: Grammatical Error Correction and Fluency Corpus for the Ukrainian Language
Provide a detailed description of the following dataset: UA-GEC
InFashAI
AI algorithms, and in particular Machine Learning (ML) algorithms, learn from data tasks that have been traditionally done by humans such as: image classification, facial recognition, linguistic translation etc. To have a good generalization capability, AI algorithms must learn from sufficiently representative data, wh...
Provide a detailed description of the following dataset: InFashAI
PNPB dataset
The dataset consists of a total of 20 videos, each of which is 5.5 minutes long in duration. The videos are captured at a resolution of 1024x1024 and at 30 frames per second. Each video contains only one pig performing the Novel Object Recognition task. It contains annotations for the following tasks: *Action Recogni...
Provide a detailed description of the following dataset: PNPB dataset
Place Pulse 2.0
Place Pulse is a crowdsourcing effort that aims to map which areas of a city are perceived as safer, livelier, wealthier, more active, beautiful and friendly. By asking users to select images from a pair, Place Pulse collected more than 1.5 million reports that evaluate more than 100,000 images from 56 cities.
Provide a detailed description of the following dataset: Place Pulse 2.0
Russian Event2Mind
The work provides a comprehensive overview of the corpus for the Russian language for the commonsense inference task. Namely, we construct event phrases, which cover a wide range of everyday situations with labelled intents and reactions of the event main participant and emotions of other people involved. Example: ...
Provide a detailed description of the following dataset: Russian Event2Mind
Taiga Corpus
Taiga is a corpus, where text sources and their meta-information are collected according to popular ML tasks. Each text in corpus is represented in plain text and with morphological and syntactic annotation (UDPipe, homonymy resolved automatically) + has metainformation - date, theme, authorship, text difficulcy…etc...
Provide a detailed description of the following dataset: Taiga Corpus
Morph Call
Morph Call is a suite of 46 probing tasks for four Indo-European languages that fall under different morphology: Russian, French, English, and German. The tasks are designed to explore the morphosyntactic content of multilingual transformers which is a less studied aspect at the moment. The tasks are divided into fo...
Provide a detailed description of the following dataset: Morph Call
UDIS-D
UDIS-D is a large image dataset for image stitching or image registration. It contains different overlap rates, varying degrees of parallax, and variable scenes such as indoor, outdoor, night, dark, snow, and zooming.
Provide a detailed description of the following dataset: UDIS-D
BCR dataset
Blender Cycles Ray-tracing (BCR) dataset contains 2449 high-quality images rendered from 1463 models. We render the images at a range of spp rates, including 1-8, 12, 16, 32, 64, 250, 1000, and 4000 spp. All the images are rendered at the resolution of 1080p. Each image contains not only the final rendered result but a...
Provide a detailed description of the following dataset: BCR dataset
RuCoS
Russian reading comprehension with Commonsense reasoning (RuCoS) is a large-scale reading comprehension dataset that requires commonsense reasoning. RuCoS consists of queries automatically generated from CNN/Daily Mail news articles; the answer to each query is a text span from a summarizing passage of the correspondin...
Provide a detailed description of the following dataset: RuCoS
DaNetQA
DaNetQA is a question answering dataset for yes/no questions. These questions are naturally occurring ---they are generated in unprompted and unconstrained settings. Each example is a triplet of (question, passage, answer), with the title of the page as optional additional context. The text-pair classification setup...
Provide a detailed description of the following dataset: DaNetQA
RWSD
A Winograd schema is a pair of sentences that differ in only one or two words and that contain an ambiguity that is resolved in opposite ways in the two sentences and requires the use of world knowledge and reasoning for its resolution. The schema takes its name from a well-known example by Terry Winograd. The set w...
Provide a detailed description of the following dataset: RWSD
RUSSE
WiC: The Word-in-Context Dataset A reliable benchmark for the evaluation of context-sensitive word embeddings. Depending on its context, an ambiguous word can refer to multiple, potentially unrelated, meanings. Mainstream static word embeddings, such as Word2vec and GloVe, are unable to reflect this dynamic semantic...
Provide a detailed description of the following dataset: RUSSE
TERRa
Textual Entailment Recognition has been proposed recently as a generic task that captures major semantic inference needs across many NLP applications, such as Question Answering, Information Retrieval, Information Extraction, and Text Summarization. This task requires to recognize, given two text fragments, whether the...
Provide a detailed description of the following dataset: TERRa
MuSeRC
We present a reading comprehension challenge in which questions can only be answered by taking into account information from multiple sentences. The dataset is the first to study multi-sentence inference at scale, with an open-ended set of question types that requires reasoning skills. ### Task Type Binary classifi...
Provide a detailed description of the following dataset: MuSeRC
PARus
Choice of Plausible Alternatives for Russian language (PARus) evaluation provides researchers with a tool for assessing progress in open-domain commonsense causal reasoning. Each question in PARus is composed of a premise and two alternatives, where the task is to select the alternative that more plausibly has a causal...
Provide a detailed description of the following dataset: PARus
RCB
The Russian Commitment Bank is a corpus of naturally occurring discourses whose final sentence contains a clause-embedding predicate under an entailment cancelling operator (question, modal, negation, antecedent of conditional). ### Task Type RTE (Recognizing Textual Entailment) Sentence Pair Classification - Entai...
Provide a detailed description of the following dataset: RCB
LiDiRus
LiDiRus is a diagnostic dataset that covers a large volume of linguistic phenomena, while allowing you to evaluate information systems on a simple test of textual entailment recognition. See more details diagnostics. ### Task Type RTE (Recognizing Textual Entailment) Sentence Pair Classification - Entailment - Not ...
Provide a detailed description of the following dataset: LiDiRus
AGORA
AGORA is a synthetic human dataset with high realism and accurate ground truth. It consists of around 14K training and 3K test images by rendering between 5 and 15 people per image using either image-based lighting or rendered 3D environments, taking care to make the images physically plausible and photoreal. In total,...
Provide a detailed description of the following dataset: AGORA
synthetic_dataset.h5
The synethetic dataset (10000 pairs of images and region, 2.95GB) is shared with the code (hdf5 dataset format).
Provide a detailed description of the following dataset: synthetic_dataset.h5
Fast Linking Numbers of Loopy Structures Dataset
Copyright (C) 2021 Ante Qu <antequ@cs.stanford.edu>. This is the dataset for this paper: Ante Qu and Doug L. James. 2021. Fast Linking Numbers for Topology Verification of Loopy Structures. ACM Trans. Graph. 40, 4, Article 106 (August 2021), 19 pages. https://doi.org/10.1145/3450626.3459778 In particular, t...
Provide a detailed description of the following dataset: Fast Linking Numbers of Loopy Structures Dataset
AIT-QA
**AIT-QA** is a dataset for Table Question Answering (Table-QA) which is specific to the airline industry. The dataset consists of 515 questions authored by human annotators on 116 tables extracted from public U.S. SEC filings of major airline companies for the fiscal years 2017-2019. It also contains annotations perta...
Provide a detailed description of the following dataset: AIT-QA
Goal
**Goal** is a novel dataset of football (or 'soccer') highlights videos with transcribed live commentaries in English. As the course of a game is unpredictable, so are commentaries, which makes them a unique resource to investigate dynamic language grounding.
Provide a detailed description of the following dataset: Goal
JetNet
JetNet is a particle cloud dataset, containing gluon, top quark, light quark jets saved in .csv format.
Provide a detailed description of the following dataset: JetNet
GPLA-12
**GPLA-12** is a new acoustic leakage dataset of gas pipelines involving 12 categories over 684 training/testing acoustic signals. The acoustic leakage signals were collected on the basis of an intact gas pipe system with external artificial leakages, and then preprocessed with structured tailoring which are turned int...
Provide a detailed description of the following dataset: GPLA-12
XAI-Bench
**XAI-Bench** is a suite of synthetic datasets along with a library for benchmarking feature attribution algorithms. Unlike real-world datasets, synthetic datasets allow the efficient computation of conditional expected values that are needed to evaluate ground-truth Shapley values and other metrics. The synthetic data...
Provide a detailed description of the following dataset: XAI-Bench
riboflavin
The dataset contains 71 samples with (normalized) expression data for 4,088 genes. The response variable is the riboflavin production rate in Bacilluss subtilis. It may be used to construct a graphical model. Introduced by Buhlmann, P., Kalisch, M., and Meier, L. (2014). High-dimensional statistics with a view towa...
Provide a detailed description of the following dataset: riboflavin
RuShiftEval
**RuShiftEval** is a manually annotated lexical semantic change dataset for Russian. Its novelty is ensured by a single set of target words annotated for their diachronic semantic shifts across three time periods, while the previous work either used only two time periods, or different sets of target words.
Provide a detailed description of the following dataset: RuShiftEval
LIVE Livestream
**LIVE Livestream** is a database for Video Quality Assessment (VQA), specifically designed for live streaming VQA research. The dataset is called the Laboratory for Image and Video Engineering (LIVE) Live stream Database. The LIVE Livestream Database includes 315 videos of 45 contents impaired by 6 types of distortion...
Provide a detailed description of the following dataset: LIVE Livestream
X-CSQA
**X-CSQA** is a multilingual dataset for Commonsense reasoning research, based on [CSQA](csqa).
Provide a detailed description of the following dataset: X-CSQA
DDPM
The Deception Detection and Physiological Monitoring (**DDPM**) dataset captures an interview scenario in which the interviewee attempts to deceive the interviewer on selected responses. The interviewee is recorded in RGB, near-infrared, and long-wave infrared, along with cardiac pulse, blood oxygenation, and audio. Af...
Provide a detailed description of the following dataset: DDPM
rSoccer
**rSoccer** is an open-source simulator for the IEEE Very Small Size Soccer and the Small Size League optimized for reinforcement learning experiments.
Provide a detailed description of the following dataset: rSoccer
VideoMatting108
**VideoMatting108** is a large-scale video matting and trimap generation dataset with 80 training and 28 validation foreground video clips with ground-truth alpha mattes.
Provide a detailed description of the following dataset: VideoMatting108
Vāksañcayaḥ
This Sanskrit speech corpus has more than 78 hours of audio data and contains recordings of 45,953 sentences with a sampling rate of 22KHz. The content is mainly readings of texts spanning over various Śāstras of Saṃskṛtam literature and also includes contemporary stories, radio program, extempore discourse, etc.
Provide a detailed description of the following dataset: Vāksañcayaḥ
ADNI
Alzheimer's Disease Neuroimaging Initiative (ADNI) is a multisite study that aims to improve clinical trials for the prevention and treatment of Alzheimer’s disease (AD).[1] This cooperative study combines expertise and funding from the private and public sector to study subjects with AD, as well as those who may devel...
Provide a detailed description of the following dataset: ADNI
Ambiguous-HOI
Ambiguous-HOI is a challenging dataset containing ambiguous human-object interaction images for HOI detection based on HICO-DET.
Provide a detailed description of the following dataset: Ambiguous-HOI