dataset_name stringlengths 2 128 | description stringlengths 1 9.7k | prompt stringlengths 59 185 |
|---|---|---|
SG-NLG | The SG-NLG dataset is a pre-processed version of the [DSTC8 Schema-Guided Dialogue SGD dataset](https://paperswithcode.com/dataset/sgd), designed specifically for data-to-text Natural Language Generation (NLG). The original DSTC8 SGD contains ~20,000 dialogues spanning across ~20 domains.
This SG-NLG dataset is desi... | Provide a detailed description of the following dataset: SG-NLG |
ShapenetRender | **ShapenetRender**er is an extension of the ShapeNet Core dataset which has more variation in camera angles. For each mesh model, the dataset provides 36 views with smaller variation and 36 views with larger variation. The resolution of the newly rendered images is 224x224 in contrast to the 137x137 original resolution... | Provide a detailed description of the following dataset: ShapenetRender |
ShapeNet-Skeleton | The **ShapeNet-Skeleton** dataset has ground-truth skeleton point sets and skeletal volumes for object instances in the ShapeNet dataset.
Source: [https://arxiv.org/pdf/2008.05742.pdf](https://arxiv.org/pdf/2008.05742.pdf) | Provide a detailed description of the following dataset: ShapeNet-Skeleton |
3D Shapes Dataset | 3dshapes is a dataset of 3D shapes procedurally generated from 6 ground truth independent latent factors. These factors are floor colour, wall colour, object colour, scale, shape and orientation. | Provide a detailed description of the following dataset: 3D Shapes Dataset |
ShapeStacks | A simulation-based dataset featuring 20,000 stack configurations composed of a variety of elementary geometric primitives richly annotated regarding semantics and structural stability. | Provide a detailed description of the following dataset: ShapeStacks |
ShEMO | The database includes 3000 semi-natural utterances, equivalent to 3 hours and 25 minutes of speech data extracted from online radio plays. The ShEMO covers speech samples of 87 native-Persian speakers for five basic emotions including anger, fear, happiness, sadness and surprise, as well as neutral state. | Provide a detailed description of the following dataset: ShEMO |
SHIDC-BC-Ki-67 | Benchmark for BC Ki-67 stained cell detection and further annotated classification of cells. | Provide a detailed description of the following dataset: SHIDC-BC-Ki-67 |
ShopSign | A newly developed natural scene text dataset of Chinese shop signs in street views. | Provide a detailed description of the following dataset: ShopSign |
SIDD | SIDD is an image denoising dataset containing 30,000 noisy images from 10 scenes under different lighting conditions using five representative smartphone cameras. Ground truth images are provided along with the noisy images. | Provide a detailed description of the following dataset: SIDD |
SidechainNet | **SidechainNet** is a protein structure prediction dataset that directly extends ProteinNet. Specifically, SidechainNet adds measurements for protein angles and coordinates that describe the complete, all-atom protein structure (backbone and sidechain, excluding hydrogens) instead of the protein backbone alone.
Source... | Provide a detailed description of the following dataset: SidechainNet |
SIDOD | SIDOD is a new, publicly-available image dataset generated by the NVIDIA Deep Learning Data Synthesizer intended for use in object detection, pose estimation, and tracking applications. This dataset contains 144k stereo image pairs that synthetically combine 18 camera viewpoints of three photorealistic virtual environm... | Provide a detailed description of the following dataset: SIDOD |
Simitate | **Simitate** is a hybrid benchmarking suite targeting the evaluation of approaches for imitation learning. It consists on a dataset containing 1938 sequences where humans perform daily activities in a realistic environment. The dataset is strongly coupled with an integration into a simulator. RGB and depth streams with... | Provide a detailed description of the following dataset: Simitate |
SIMMC | Situated Interactive MultiModal Conversations (**SIMMC**) is the task of taking multimodal actions grounded in a co-evolving multimodal input content in addition to the dialog history. This dataset contains two SIMMC datasets totalling ~13K human-human dialogs (~169K utterances) using a multimodal Wizard-of-Oz (WoZ) se... | Provide a detailed description of the following dataset: SIMMC |
simply-CLEVR | The **simply-CLEVR** dataset aims to provide a benchmark dataset that can be used for transparent quantitative evaluation of explanation methods (aka heatmaps/XAI methods).
It is made of simple Visual Question Answering (VQA) questions, which are derived from the original CLEVR task, and where each question is accompan... | Provide a detailed description of the following dataset: simply-CLEVR |
SIS | Comprises of 400 naturalistic usages of literature-informed verbs spanning the spectrum of symmetry-asymmetry. | Provide a detailed description of the following dataset: SIS |
SI-SCORE | A synthetic dataset uses for a systematic analysis across common factors of variation. | Provide a detailed description of the following dataset: SI-SCORE |
SIZER | Dataset of clothing size variation which includes different subjects wearing casual clothing items in various sizes, totaling to approximately 2000 scans. This dataset includes the scans, registrations to the SMPL model, scans segmented in clothing parts, garment category and size labels. | Provide a detailed description of the following dataset: SIZER |
SketchGraphs | **SketchGraphs** is a dataset of 15 million sketches extracted from real-world CAD models intended to facilitate research in both ML-aided design and geometric program induction.
Each sketch is represented as a geometric constraint graph where edges denote designer-imposed geometric relationships between primitives, th... | Provide a detailed description of the following dataset: SketchGraphs |
ShoeV2 | **ShoeV2** is a dataset of 2,000 photos and 6648 sketches of shoes. The dataset is designed for fine-grained sketch-based image retrieval. | Provide a detailed description of the following dataset: ShoeV2 |
Skill2vec | Collects a huge number of job descriptions from Dice.com - one of the most popular career website about Tech jobs in USA. From these job descriptions, skills are extracted for each one by using skills dictionary. Now, the dataset is presented by a list of collections of skills based on job descriptions. After crawling,... | Provide a detailed description of the following dataset: Skill2vec |
SKU110K-R | **SKU110K-R** is a dataset relabeled with oriented bounding boxes based on SKU110K. It is focused on evaluating oriented and densely packed object detection.
Source: [https://github.com/Anymake/DRN_CVPR2020](https://github.com/Anymake/DRN_CVPR2020)
Image Source: [https://github.com/Anymake/DRN_CVPR2020](https://github... | Provide a detailed description of the following dataset: SKU110K-R |
SlowFlow | **SlowFlow** is an optical flow dataset collected by applying Slow Flow technique on data from a high-speed camera and analyzing the performance of the state-of-the-art in optical flow under various levels of motion blur. | Provide a detailed description of the following dataset: SlowFlow |
SLURP | A new challenging dataset in English spanning 18 domains, which is substantially bigger and linguistically more diverse than existing datasets. | Provide a detailed description of the following dataset: SLURP |
SMHD | A novel large dataset of social media posts from users with one or multiple mental health conditions along with matched control users. | Provide a detailed description of the following dataset: SMHD |
SmokEng | SmokEng is a dataset of 3144 tweets, which are selected based on the presence of colloquial slang related to smoking and analyze it based on the semantics of the tweet. | Provide a detailed description of the following dataset: SmokEng |
SMS Spam Collection Data Set | This corpus has been collected from free or free for research sources at the Internet:
- A collection of 425 SMS spam messages was manually extracted from the Grumbletext Web site. This is a UK forum in which cell phone users make public claims about SMS spam messages, most of them without reporting the very spam me... | Provide a detailed description of the following dataset: SMS Spam Collection Data Set |
SMS-WSJ | Spatialized Multi-Speaker Wall Street Journal (SMS-WSJ) consists of artificially mixed speech taken from the WSJ database, but unlike earlier databases this one considers all WSJ0+1 utterances and takes care of strictly separating the speaker sets present in the training, validation and test sets. | Provide a detailed description of the following dataset: SMS-WSJ |
SNLI-VE | Visual Entailment (VE) consists of image-sentence pairs whereby a premise is defined by an image, rather than a natural language sentence as in traditional Textual Entailment tasks. The goal of a trained VE model is to predict whether the image semantically entails the text. **SNLI-VE** is a dataset for VE which is bas... | Provide a detailed description of the following dataset: SNLI-VE |
So2Sat LCZ42 | So2Sat LCZ42 consists of local climate zone (LCZ) labels of about half a million Sentinel-1 and Sentinel-2 image patches in 42 urban agglomerations (plus 10 additional smaller areas) across the globe. This dataset was labeled by 15 domain experts following a carefully designed labeling work flow and evaluation process ... | Provide a detailed description of the following dataset: So2Sat LCZ42 |
SOBA | A new dataset called SOBA, named after Shadow-OBject Association, with 3,623 pairs of shadow and object instances in 1,000 photos, each with individual labeled masks. | Provide a detailed description of the following dataset: SOBA |
SoccerData | A dataset of 4562 images of which 4152 images contain a soccer ball. | Provide a detailed description of the following dataset: SoccerData |
SoccerDB | Comprises of 171,191 video segments from 346 high-quality soccer games. The database contains 702,096 bounding boxes, 37,709 essential event labels with time boundary and 17,115 highlight annotations for object detection, action recognition, temporal action localization, and highlight detection tasks. | Provide a detailed description of the following dataset: SoccerDB |
SoccerNet | A benchmark for action spotting in soccer videos. The dataset is composed of 500 complete soccer games from six main European leagues, covering three seasons from 2014 to 2017 and a total duration of 764 hours. A total of 6,637 temporal annotations are automatically parsed from online match reports at a one minute reso... | Provide a detailed description of the following dataset: SoccerNet |
SoccerNet-v2 | A novel large-scale corpus of manual annotations for the SoccerNet video dataset, along with open challenges to encourage more research in soccer understanding and broadcast production. | Provide a detailed description of the following dataset: SoccerNet-v2 |
Social-IQ | Social-IQ is an unconstrained benchmark specifically designed to train and evaluate socially intelligent technologies. By providing a rich source of open-ended questions and answers, Social-IQ opens the door to explainable social intelligence. The dataset contains rigorously annotated and validated videos, questions an... | Provide a detailed description of the following dataset: Social-IQ |
SMM4H | Social Media Mining for Health (SMM4H) Shared Task is a massive data source for biomedical and public health applications. | Provide a detailed description of the following dataset: SMM4H |
SoloDance | A large-scale HVMT dataset named SoloDance. | Provide a detailed description of the following dataset: SoloDance |
Some Like it Hoax | **Some Like it Hoax** is a fake news detection dataset consisting of 15,500 Facebook posts and 909,236 users. | Provide a detailed description of the following dataset: Some Like it Hoax |
SONYC-UST-V2 | A dataset for urban sound tagging with spatiotemporal information. This dataset is aimed for the development and evaluation of machine listening systems for real-world urban noise monitoring. While datasets of urban recordings are available, this dataset provides the opportunity to investigate how spatiotemporal metada... | Provide a detailed description of the following dataset: SONYC-UST-V2 |
SoyCultivarVein | The SoyCultivarVein dataset is a publicly available dataset, which comprises 100 categories (cultivars) with 6 samples (leaf images) in each cultivar and thus has a total number of 100×6 = 600 images (Yu et al. 2019). The leaves in the SoyCultivarVein dataset are highly similar due to the fact that they all belong to t... | Provide a detailed description of the following dataset: SoyCultivarVein |
SP-10K | A large-scale evaluation set that provides human ratings for the plausibility of 10,000 SP pairs over five SP relations, covering 2,500 most frequent verbs, nouns, and adjectives in American English. | Provide a detailed description of the following dataset: SP-10K |
SpaceNet MVOI | An open source Multi-View Overhead Imagery dataset with 27 unique looks from a broad range of viewing angles (-32.5 degrees to 54.0 degrees). Each of these images cover the same 665 square km geographic extent and are annotated with 126,747 building footprint labels, enabling direct assessment of the impact of viewpoin... | Provide a detailed description of the following dataset: SpaceNet MVOI |
Spaceship Dataset | The Spaceship dataset is a dataset for evaluating agents’ ability to learn to solve a class of physics-based tasks. The tasks consist on a spaceship that has to reach a the mothership in 11 steps, in an environment where static planets exert gravitational forces on the spaceship, which induce complex non-linear dynamic... | Provide a detailed description of the following dataset: Spaceship Dataset |
SPair-71k | SPair-71k contains 70,958 image pairs with diverse variations in viewpoint and scale. Compared to previous datasets, it is significantly larger in number and contains more accurate and richer annotations. | Provide a detailed description of the following dataset: SPair-71k |
SPARE3D | Contains three types of 2D-3D reasoning tasks on view consistency, camera pose, and shape generation, with increasing difficulty. | Provide a detailed description of the following dataset: SPARE3D |
SpatialSense Benchmark | SpatialSense Benchmark is a dataset specializing in spatial relation recognition which captures a broad spectrum of such challenges, allowing for proper benchmarking of computer vision techniques. | Provide a detailed description of the following dataset: SpatialSense Benchmark |
SpeakingFaces | SpeakingFaces is a publicly-available large-scale dataset developed to support multimodal machine learning research in contexts that utilize a combination of thermal, visual, and audio data streams; examples include human-computer interaction (HCI), biometric authentication, recognition systems, domain transfer, and sp... | Provide a detailed description of the following dataset: SpeakingFaces |
SpectroVision | **SpectroVision** is a dataset of 14,400 high resolution texture images and spectral measurements collected from a PR2 mobile manipulator that interacted with 144 household objects from eight material categories.
Source: [https://github.com/Healthcare-Robotics/spectrovision](https://github.com/Healthcare-Robotics/spec... | Provide a detailed description of the following dataset: SpectroVision |
SPEECH-COCO | SPEECH-COCO contains speech captions that are generated using text-to-speech (TTS) synthesis resulting in 616,767 spoken captions (more than 600h) paired with images. | Provide a detailed description of the following dataset: SPEECH-COCO |
SPIRS | A first-of-its-kind large dataset of sarcastic/non-sarcastic tweets with high-quality labels and extra features: (1) sarcasm perspective labels (2) new contextual features. The dataset is expected to advance sarcasm detection research. | Provide a detailed description of the following dataset: SPIRS |
SPLASH | A dataset of utterances, incorrect SQL interpretations and the corresponding natural language feedback. | Provide a detailed description of the following dataset: SPLASH |
Spoken-SQuAD | In SpokenSQuAD, the document is in spoken form, the input question is in the form of text and the answer to each question is always a span in the document. The following procedures were used to generate spoken documents from the original SQuAD dataset. First, the Google text-to-speech system was used to generate the sp... | Provide a detailed description of the following dataset: Spoken-SQuAD |
Spotify Podcast | A set of approximately 100K podcast episodes comprised of raw audio files along with accompanying ASR transcripts. This represents over 47,000 hours of transcribed audio, and is an order of magnitude larger than previous speech-to-text corpora. | Provide a detailed description of the following dataset: Spotify Podcast |
SQuAD-es | Stanford Question Answering Dataset (SQuAD) into Spanish. | Provide a detailed description of the following dataset: SQuAD-es |
SQuAD-it | SQuAD-it is derived from the SQuAD dataset and it is obtained through semi-automatic translation of the SQuAD dataset into Italian. It represents a large-scale dataset for open question answering processes on factoid questions in Italian. The dataset contains more than 60,000 question/answer pairs derived from the orig... | Provide a detailed description of the following dataset: SQuAD-it |
SQuAD-shifts | Provides four new test sets for the Stanford Question Answering Dataset (SQuAD) and evaluate the ability of question-answering systems to generalize to new data. | Provide a detailed description of the following dataset: SQuAD-shifts |
SQUID | A dataset of images taken in different locations with varying water properties, showing color charts in the scenes. Moreover, to obtain ground truth, the 3D structure of the scene was calculated based on stereo imaging. This dataset enables a quantitative evaluation of restoration algorithms on natural images. | Provide a detailed description of the following dataset: SQUID |
STAIR Actions Captions | A large-scale Japanese video caption dataset consisting of 79,822 videos and 399,233 captions. Each caption in the dataset describes a video in the form of "who does what and where." | Provide a detailed description of the following dataset: STAIR Actions Captions |
Standardized Project Gutenberg Corpus | The **Standardized Project Gutenberg Corpus** (SPGC) is an open science approach to a curated version of the complete PG data containing more than 50,000 books and more than 3×109 word-tokens.
Source: [https://arxiv.org/abs/1812.08092](https://arxiv.org/abs/1812.08092) | Provide a detailed description of the following dataset: Standardized Project Gutenberg Corpus |
StanfordExtra | An 'in the wild' dataset of 20,580 dog images for which 2D joint and silhouette annotations were collected. | Provide a detailed description of the following dataset: StanfordExtra |
StaQC | **StaQC** (Stack Overflow Question-Code pairs) is a large dataset of around 148K Python and 120K SQL domain question-code pairs, which are automatically mined from StackOverflow. | Provide a detailed description of the following dataset: StaQC |
STAR | A schema-guided task-oriented dialog dataset consisting of 127,833 utterances and knowledge base queries across 5,820 task-oriented dialogs in 13 domains that is especially designed to facilitate task and domain transfer learning in task-oriented dialog. | Provide a detailed description of the following dataset: STAR |
StereoMSI | StereoMSI comprises of 350 registered colour-spectral image pairs. The dataset has been used for the two tracks of the PIRM2018 challenge. | Provide a detailed description of the following dataset: StereoMSI |
stickerchart | The Stickerchat dataset is a large-scale real-world dialog dataset with stickers which contains 340K multi-turn dialog and sticker pairs.
Source: [https://arxiv.org/abs/2003.04679](https://arxiv.org/abs/2003.04679) | Provide a detailed description of the following dataset: stickerchart |
Store dataset | The Store Dataset is a dataset for estimating 3D poses of multiple humans in real-time. It is captured inside two kinds of simulated stores with 12 and 28 cameras, respectively.
Source: [https://arxiv.org/abs/2003.03972](https://arxiv.org/abs/2003.03972) | Provide a detailed description of the following dataset: Store dataset |
Story Commonsense | Story Commonsense is a new large-scale dataset with rich low-level annotations and establishes baseline performance on several new tasks, suggesting avenues for future research. | Provide a detailed description of the following dataset: Story Commonsense |
Stream-51 | A new dataset for streaming classification consisting of temporally correlated images from 51 distinct object categories and additional evaluation classes outside of the training distribution to test novelty recognition. | Provide a detailed description of the following dataset: Stream-51 |
Street Dataset | A real-world image dataset that contains more than 900 images generated from 26 street cameras and 7 object categories annotated with detailed bounding box. The data distribution is non-IID and unbalanced, reflecting the characteristic real-world federated learning scenarios. | Provide a detailed description of the following dataset: Street Dataset |
Exact Street2Shop | A dataset containing 404,683 shop photos collected from 25 different online retailers and 20,357 street photos, providing a total of 39,479 clothing item matches between street and shop photos. | Provide a detailed description of the following dataset: Exact Street2Shop |
StreetHazards | StreetHazards is a synthetic dataset for anomaly detection, created by inserting a diverse array of foreign objects into driving scenes and re-render the scenes with these novel objects. | Provide a detailed description of the following dataset: StreetHazards |
StreetLearn | An interactive, first-person, partially-observed visual environment that uses Google Street View for its photographic content and broad coverage, and give performance baselines for a challenging goal-driven navigation task. | Provide a detailed description of the following dataset: StreetLearn |
Street View Image, Pose, and 3D Cities Dataset | A large-scale dataset composed of object-centric street view scenes along with point correspondences and camera pose information. | Provide a detailed description of the following dataset: Street View Image, Pose, and 3D Cities Dataset |
Structured3D | **Structured3D** is a large-scale photo-realistic dataset containing 3.5K house designs (a) created by professional designers with a variety of ground truth 3D structure annotations (b) and generate photo-realistic 2D images (c).
The dataset consists of rendering images and corresponding ground truth annotations (e.g.... | Provide a detailed description of the following dataset: Structured3D |
ST-VQA | ST-VQA aims to highlight the importance of exploiting high-level semantic information present in images as textual cues in the VQA process. | Provide a detailed description of the following dataset: ST-VQA |
SubEdits | **SubEdits** is a human-annnoated post-editing dataset of neural machine translation outputs, compiled from in-house NMT outputs and human post-edits of subtitles form Rakuten Viki. It is collected from English-German annotations and contains 160k triplets. | Provide a detailed description of the following dataset: SubEdits |
SubjQA | **SubjQA** is a question answering dataset that focuses on subjective (as opposed to factual) questions and answers. The dataset consists of roughly 10,000 questions over reviews from 6 different domains: books, movies, grocery, electronics, TripAdvisor (i.e. hotels), and restaurants. Each question is paired with a rev... | Provide a detailed description of the following dataset: SubjQA |
Surveillance Camera Fight Dataset | The dataset is collected from the Youtube videos that contains fight instances in it. Also, some non-fight sequences from regular surveillance camera videos are included.
* There are 300 videos in total as 150 fight + 150 non-fight
* Videos are 2-second long
* Only the fight related parts are included in the samples
S... | Provide a detailed description of the following dataset: Surveillance Camera Fight Dataset |
SuspectGuilt Corpus | A corpus of annotated crime stories from English-language newspapers in the U.S. For SuspectGuilt, annotators read short crime articles and provided text-level ratings concerning the guilt of the main suspect as well as span-level annotations indicating which parts of the story they felt most influenced their ratings. ... | Provide a detailed description of the following dataset: SuspectGuilt Corpus |
SVD | SVD is a large-scale short video dataset, which contains over 500,000 short videos collected from http://www.douyin.com and over 30,000 labeled pairs of near-duplicate videos. | Provide a detailed description of the following dataset: SVD |
SVIRO | Contains bounding boxes for object detection, instance segmentation masks, keypoints for pose estimation and depth images for each synthetic scenery as well as images for each individual seat for classification. | Provide a detailed description of the following dataset: SVIRO |
SWAX | Comprised of real human and wax figure images and videos that endorse the problem of face spoofing detection. The dataset consists of more than 1800 face images and 110 videos of 55 people/waxworks, arranged in training, validation and test sets with a large range in expression, illumination and pose variations. | Provide a detailed description of the following dataset: SWAX |
SweetRS | Uses a platform with 77 candies and sweets to rank. Over 2000 users submitted over 44000 grades resulting in a matrix with 28% coverage. | Provide a detailed description of the following dataset: SweetRS |
Swiss3DCities | Swiss3DCities is a dataset that is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras. | Provide a detailed description of the following dataset: Swiss3DCities |
Synscapes | Synscapes is a synthetic dataset for street scene parsing created using photorealistic rendering techniques, and show state-of-the-art results for training and validation as well as new types of analysis. | Provide a detailed description of the following dataset: Synscapes |
SynthCity | **SynthCity** is a 367.9M point synthetic full colour Mobile Laser Scanning point cloud. Every point is assigned a label from one of nine categories. | Provide a detailed description of the following dataset: SynthCity |
Synthetic Human Model Dataset | A synthetic dataset for evaluating non-rigid 3D human reconstruction based on conventional RGB-D cameras. The dataset consist of seven motion sequences of a single human model. | Provide a detailed description of the following dataset: Synthetic Human Model Dataset |
Synthetic Keystroke | This dataset is a large-scale synthetic dataset to simulate the attack scenario for a keystroke inference attack.
Source: [https://arxiv.org/abs/2009.05796](https://arxiv.org/abs/2009.05796) | Provide a detailed description of the following dataset: Synthetic Keystroke |
SYNTHIA-AL | Specially designed to evaluate active learning for video object detection in road scenes. | Provide a detailed description of the following dataset: SYNTHIA-AL |
Synthinel-1 | **Synthinel-1** is a collection of synthetic overhead imagery with full pixel-wise building segmentation labels.
Source: [https://github.com/timqqt/Synthinel](https://github.com/timqqt/Synthinel)
Image Source: [https://github.com/timqqt/Synthinel](https://github.com/timqqt/Synthinel) | Provide a detailed description of the following dataset: Synthinel-1 |
SYSU-30k | **SYSU-30k** contains 30k categories of persons, which is about 20 times larger than CUHK03 (1.3k categories) and Market1501 (1.5k categories), and 30 times larger than ImageNet (1k categories). SYSU-30k contains 29,606,918 images. Moreover, SYSU-30k provides not only a large platform for the weakly supervised ReID pro... | Provide a detailed description of the following dataset: SYSU-30k |
SYSU-CEUS | The **SYSU-CEUS** dataset consists of three types of Focal liver lesions (FLLs): 186 HCC instances, 109 HEM instances and 58 FNH instances (i.e.,186 malignant instances and 167 benign instances).
This dataset is collected from the First Affiliated Hospital, Sun Yat-sen University. The equipment used was Aplio SSA-770A ... | Provide a detailed description of the following dataset: SYSU-CEUS |
TACO | **TACO** is a growing image dataset of waste in the wild. It contains images of litter taken under diverse environments: woods, roads and beaches. These images are manually labelled and segmented according to a hierarchical taxonomy to train and evaluate object detection algorithms. The annotations are provided in COCO... | Provide a detailed description of the following dataset: TACO |
TACoS Multi-Level Corpus | Augments the video-description dataset TACoS with short and single sentence descriptions. | Provide a detailed description of the following dataset: TACoS Multi-Level Corpus |
Talk2Car | The **Talk2Car** dataset finds itself at the intersection of various research domains, promoting the development of cross-disciplinary solutions for improving the state-of-the-art in grounding natural language into visual space. The annotations were gathered with the following aspects in mind:
Free-form high quality n... | Provide a detailed description of the following dataset: Talk2Car |
Talk2Nav | Talk2Nav is a large-scale dataset with verbal navigation instructions. | Provide a detailed description of the following dataset: Talk2Nav |
TalkDown | **TalkDown** is a labelled dataset for condescension detection in context. The dataset is derived from Reddit, a set of online communities that is diverse in content and tone. The dataset is built from COMMENT and REPLY pairs in which the REPLY targets a specific quoted span (QUOTED) in the COMMENT as being condescendi... | Provide a detailed description of the following dataset: TalkDown |
Talk the Walk | Talk The Walk is a large-scale dialogue dataset grounded in
action and perception. The task involves two agents (a “guide” and a “tourist”)
that communicate via natural language in order to achieve a common goal: having
the tourist navigate to a given target location. | Provide a detailed description of the following dataset: Talk the Walk |
TAO | TAO is a federated dataset for Tracking Any Object, containing 2,907 high resolution videos, captured in diverse environments, which are half a minute long on average. A bottom-up approach was used for discovering a large vocabulary of 833 categories, an order of magnitude more than prior tracking benchmarks.
The d... | Provide a detailed description of the following dataset: TAO |
TaoDescribe | The **TaoDescribe** dataset contains 2,129,187 product titles and descriptions in Chinese.
Source: [https://github.com/qibinc/KOBE](https://github.com/qibinc/KOBE) | Provide a detailed description of the following dataset: TaoDescribe |
TaPaCo | TaPaCo is a freely available paraphrase corpus for 73 languages extracted from the Tatoeba database. | Provide a detailed description of the following dataset: TaPaCo |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.