url
stringlengths
58
61
repository_url
stringclasses
1 value
labels_url
stringlengths
72
75
comments_url
stringlengths
67
70
events_url
stringlengths
65
68
html_url
stringlengths
46
51
id
int64
599M
1.32B
node_id
stringlengths
18
32
number
int64
1
4.75k
title
stringlengths
1
276
user
dict
labels
list
state
stringclasses
2 values
locked
bool
1 class
assignee
dict
assignees
list
milestone
dict
comments
list
created_at
int64
1,587B
1,659B
updated_at
int64
1,587B
1,659B
closed_at
int64
1,587B
1,659B
author_association
stringclasses
3 values
active_lock_reason
null
draft
bool
2 classes
pull_request
dict
body
stringlengths
0
228k
reactions
dict
timeline_url
stringlengths
67
70
performed_via_github_app
null
state_reason
stringclasses
3 values
is_pull_request
bool
2 classes
https://api.github.com/repos/huggingface/datasets/issues/683
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/683/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/683/comments
https://api.github.com/repos/huggingface/datasets/issues/683/events
https://github.com/huggingface/datasets/pull/683
710,942,704
MDExOlB1bGxSZXF1ZXN0NDk0NzAwNzY1
683
Fix wrong delimiter in text dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,601,372,604,000
1,620,239,071,000
1,601,372,646,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/683", "html_url": "https://github.com/huggingface/datasets/pull/683", "diff_url": "https://github.com/huggingface/datasets/pull/683.diff", "patch_url": "https://github.com/huggingface/datasets/pull/683.patch", "merged_at": null }
The delimiter is set to the bell character as it is used nowhere is text files usually. However in the text dataset the delimiter was set to `\b` which is backspace in python, while the bell character is `\a`. I replace \b by \a Hopefully it fixes issues mentioned by some users in #622
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/683/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/683/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/682
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/682/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/682/comments
https://api.github.com/repos/huggingface/datasets/issues/682/events
https://github.com/huggingface/datasets/pull/682
710,325,399
MDExOlB1bGxSZXF1ZXN0NDk0MTkzMzEw
682
Update navbar chapter titles color
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,601,303,717,000
1,601,314,213,000
1,601,314,212,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/682", "html_url": "https://github.com/huggingface/datasets/pull/682", "diff_url": "https://github.com/huggingface/datasets/pull/682.diff", "patch_url": "https://github.com/huggingface/datasets/pull/682.patch", "merged_at": 1601314212000 }
Consistency with the color change that was done in transformers at https://github.com/huggingface/transformers/pull/7423 It makes the background-color of the chapter titles in the docs navbar darker, to differentiate them from the inner sections. see changes [here](https://691-250213286-gh.circle-artifacts.com/0/do...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/682/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/682/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/681
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/681/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/681/comments
https://api.github.com/repos/huggingface/datasets/issues/681/events
https://github.com/huggingface/datasets/pull/681
710,075,721
MDExOlB1bGxSZXF1ZXN0NDkzOTkwMjEz
681
Adding missing @property (+2 small flake8 fixes).
{ "login": "Narsil", "id": 204321, "node_id": "MDQ6VXNlcjIwNDMyMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Narsil", "html_url": "https://github.com/Narsil", "followers_url": "https://api.github.com/users/Narsil/follow...
[]
closed
false
null
[]
null
[]
1,601,283,233,000
1,601,288,773,000
1,601,288,769,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/681", "html_url": "https://github.com/huggingface/datasets/pull/681", "diff_url": "https://github.com/huggingface/datasets/pull/681.diff", "patch_url": "https://github.com/huggingface/datasets/pull/681.patch", "merged_at": 1601288769000 }
Fixes #678
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/681/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/681/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/680
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/680/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/680/comments
https://api.github.com/repos/huggingface/datasets/issues/680/events
https://github.com/huggingface/datasets/pull/680
710,066,138
MDExOlB1bGxSZXF1ZXN0NDkzOTgyMjY4
680
Fix bug related to boolean in GAP dataset.
{ "login": "otakumesi", "id": 14996977, "node_id": "MDQ6VXNlcjE0OTk2OTc3", "avatar_url": "https://avatars.githubusercontent.com/u/14996977?v=4", "gravatar_id": "", "url": "https://api.github.com/users/otakumesi", "html_url": "https://github.com/otakumesi", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi !\r\n\r\nGood catch, thanks for creating this PR :)\r\n\r\nCould you also regenerate the metadata for this dataset using \r\n```\r\ndatasets-cli test ./datasets/gap --save_infos --all_configs\r\n```\r\n\r\nThat'd be awesome", "@lhoestq Thank you for your revieing!!!\r\n\r\nI've performed it and have read CONT...
1,601,282,379,000
1,601,394,887,000
1,601,394,887,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/680", "html_url": "https://github.com/huggingface/datasets/pull/680", "diff_url": "https://github.com/huggingface/datasets/pull/680.diff", "patch_url": "https://github.com/huggingface/datasets/pull/680.patch", "merged_at": 1601394887000 }
### Why I did The value in `row["A-coref"]` and `row["B-coref"]` is `'TRUE'` or `'FALSE'`. This type is `string`, then `bool('FALSE')` is equal to `True` in Python. So, both rows are transformed into `True` now. So, I modified this problem. ### What I did I modified `bool(row["A-coref"])` and `bool(row["B-cor...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/680/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/680/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/679
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/679/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/679/comments
https://api.github.com/repos/huggingface/datasets/issues/679/events
https://github.com/huggingface/datasets/pull/679
710,065,838
MDExOlB1bGxSZXF1ZXN0NDkzOTgyMDMx
679
Fix negative ids when slicing with an array
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,601,282,348,000
1,601,304,140,000
1,601,304,139,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/679", "html_url": "https://github.com/huggingface/datasets/pull/679", "diff_url": "https://github.com/huggingface/datasets/pull/679.diff", "patch_url": "https://github.com/huggingface/datasets/pull/679.patch", "merged_at": 1601304139000 }
```python from datasets import Dataset d = ds.Dataset.from_dict({"a": range(10)}) print(d[[0, -1]]) # OverflowError ``` raises an error because of the negative id. This PR fixes that. Fix #668
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/679/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/679/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/678
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/678/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/678/comments
https://api.github.com/repos/huggingface/datasets/issues/678/events
https://github.com/huggingface/datasets/issues/678
710,060,497
MDU6SXNzdWU3MTAwNjA0OTc=
678
The download instructions for c4 datasets are not contained in the error message
{ "login": "Narsil", "id": 204321, "node_id": "MDQ6VXNlcjIwNDMyMQ==", "avatar_url": "https://avatars.githubusercontent.com/u/204321?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Narsil", "html_url": "https://github.com/Narsil", "followers_url": "https://api.github.com/users/Narsil/follow...
[]
closed
false
null
[]
null
[ "Good catch !\r\nIndeed the `@property` is missing.\r\n\r\nFeel free to open a PR :)", "Also not that C4 is a dataset that needs an Apache Beam runtime to be generated.\r\nFor example Dataflow, Spark, Flink etc.\r\n\r\nUsually we generate the dataset on our side once and for all, but we haven't done it for C4 yet...
1,601,281,854,000
1,601,288,769,000
1,601,288,769,000
CONTRIBUTOR
null
null
null
The manual download instructions are not clear ```The dataset c4 with config en requires manual data. Please follow the manual download instructions: <bound method C4.manual_download_instructions of <datasets_modules.datasets.c4.830b0c218bd41fed439812c8dd19dbd4767d2a3faa385eb695cf8666c982b1b3.c4.C4 object at 0x7ff...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/678/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/678/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/677
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/677/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/677/comments
https://api.github.com/repos/huggingface/datasets/issues/677/events
https://github.com/huggingface/datasets/pull/677
710,055,239
MDExOlB1bGxSZXF1ZXN0NDkzOTczNDE3
677
Move cache dir root creation in builder's init
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,601,281,366,000
1,601,304,163,000
1,601,304,162,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/677", "html_url": "https://github.com/huggingface/datasets/pull/677", "diff_url": "https://github.com/huggingface/datasets/pull/677.diff", "patch_url": "https://github.com/huggingface/datasets/pull/677.patch", "merged_at": 1601304162000 }
We use lock files in the builder initialization but sometimes the cache directory where they're supposed to be was not created. To fix that I moved the builder's cache dir root creation in the builder's init. Fix #671
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/677/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/677/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/676
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/676/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/676/comments
https://api.github.com/repos/huggingface/datasets/issues/676/events
https://github.com/huggingface/datasets/issues/676
710,014,319
MDU6SXNzdWU3MTAwMTQzMTk=
676
train_test_split returns empty dataset item
{ "login": "mojave-pku", "id": 26648528, "node_id": "MDQ6VXNlcjI2NjQ4NTI4", "avatar_url": "https://avatars.githubusercontent.com/u/26648528?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mojave-pku", "html_url": "https://github.com/mojave-pku", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "The problem still exists after removing the cache files.", "Can you reproduce this example in a Colab so we can investigate? (or give more information on your software/hardware config)", "Thanks for reporting.\r\nI just found the issue, I'm creating a PR", "We'll do a release pretty soon to include the fix :...
1,601,277,573,000
1,602,078,393,000
1,602,077,886,000
NONE
null
null
null
I try to split my dataset by `train_test_split`, but after that the item in `train` and `test` `Dataset` is empty. The codes: ``` yelp_data = datasets.load_from_disk('/home/ssd4/huanglianzhe/test_yelp') print(yelp_data[0]) yelp_data = yelp_data.train_test_split(test_size=0.1) print(yelp_data) pri...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/676/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/676/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/675
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/675/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/675/comments
https://api.github.com/repos/huggingface/datasets/issues/675/events
https://github.com/huggingface/datasets/issues/675
709,818,725
MDU6SXNzdWU3MDk4MTg3MjU=
675
Add custom dataset to NLP?
{ "login": "timpal0l", "id": 6556710, "node_id": "MDQ6VXNlcjY1NTY3MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/6556710?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timpal0l", "html_url": "https://github.com/timpal0l", "followers_url": "https://api.github.com/users/timpa...
[]
closed
false
null
[]
null
[ "Yes you can have a look here: https://huggingface.co/docs/datasets/loading_datasets.html#csv-files", "No activity, closing" ]
1,601,241,770,000
1,603,184,929,000
1,603,184,929,000
CONTRIBUTOR
null
null
null
Is it possible to add a custom dataset such as a .csv to the NLP library? Thanks.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/675/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/675/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/674
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/674/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/674/comments
https://api.github.com/repos/huggingface/datasets/issues/674/events
https://github.com/huggingface/datasets/issues/674
709,661,006
MDU6SXNzdWU3MDk2NjEwMDY=
674
load_dataset() won't download in Windows
{ "login": "ThisDavehead", "id": 34422661, "node_id": "MDQ6VXNlcjM0NDIyNjYx", "avatar_url": "https://avatars.githubusercontent.com/u/34422661?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ThisDavehead", "html_url": "https://github.com/ThisDavehead", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "I have the same issue. Tried to download a few of them and not a single one is downloaded successfully.\r\n\r\nThis is the output:\r\n```\r\n>>> dataset = load_dataset('blended_skill_talk', split='train')\r\nUsing custom data configuration default <-- This step never ends\r\n```", "This was fixed i...
1,601,178,985,000
1,601,886,498,000
1,601,886,498,000
NONE
null
null
null
I don't know if this is just me or Windows. Maybe other Windows users can chime in if they don't have this problem. I've been trying to get some of the tutorials working on Windows, but when I use the load_dataset() function, it just stalls and the script keeps running indefinitely without downloading anything. I've wa...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/674/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/674/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/673
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/673/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/673/comments
https://api.github.com/repos/huggingface/datasets/issues/673/events
https://github.com/huggingface/datasets/issues/673
709,603,989
MDU6SXNzdWU3MDk2MDM5ODk=
673
blog_authorship_corpus crashed
{ "login": "Moshiii", "id": 7553188, "node_id": "MDQ6VXNlcjc1NTMxODg=", "avatar_url": "https://avatars.githubusercontent.com/u/7553188?v=4", "gravatar_id": "", "url": "https://api.github.com/users/Moshiii", "html_url": "https://github.com/Moshiii", "followers_url": "https://api.github.com/users/Moshiii/...
[ { "id": 2107841032, "node_id": "MDU6TGFiZWwyMTA3ODQxMDMy", "url": "https://api.github.com/repos/huggingface/datasets/labels/nlp-viewer", "name": "nlp-viewer", "color": "94203D", "default": false, "description": "" } ]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWe'll free some memory" ]
1,601,151,328,000
1,644,922,078,000
1,644,922,078,000
NONE
null
null
null
This is just to report that When I pick blog_authorship_corpus in https://huggingface.co/nlp/viewer/?dataset=blog_authorship_corpus I get this: ![image](https://user-images.githubusercontent.com/7553188/94349542-4364f300-0013-11eb-897d-b25660a449f0.png)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/673/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/673/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/672
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/672/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/672/comments
https://api.github.com/repos/huggingface/datasets/issues/672/events
https://github.com/huggingface/datasets/issues/672
709,575,527
MDU6SXNzdWU3MDk1NzU1Mjc=
672
Questions about XSUM
{ "login": "danyaljj", "id": 2441454, "node_id": "MDQ6VXNlcjI0NDE0NTQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2441454?v=4", "gravatar_id": "", "url": "https://api.github.com/users/danyaljj", "html_url": "https://github.com/danyaljj", "followers_url": "https://api.github.com/users/danya...
[]
open
false
null
[]
null
[ "We should try to regenerate the data using the official script.\r\nBut iirc that's what we used in the first place, so not sure why it didn't match in the first place.\r\n\r\nI'll let you know when the dataset is updated", "Thanks, looking forward to hearing your update on this thread. \r\n\r\nThis is a blocking...
1,601,140,584,000
1,603,185,367,000
null
CONTRIBUTOR
null
null
null
Hi there ✋ I'm looking into your `xsum` dataset and I have several questions on that. So here is how I loaded the data: ``` >>> data = datasets.load_dataset('xsum', version='1.0.1') >>> data['train'] Dataset(features: {'document': Value(dtype='string', id=None), 'summary': Value(dtype='string', id=None)}, nu...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/672/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/672/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/671
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/671/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/671/comments
https://api.github.com/repos/huggingface/datasets/issues/671/events
https://github.com/huggingface/datasets/issues/671
709,093,151
MDU6SXNzdWU3MDkwOTMxNTE=
671
[BUG] No such file or directory
{ "login": "jbragg", "id": 2238344, "node_id": "MDQ6VXNlcjIyMzgzNDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2238344?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jbragg", "html_url": "https://github.com/jbragg", "followers_url": "https://api.github.com/users/jbragg/foll...
[]
closed
false
null
[]
null
[]
1,601,051,934,000
1,601,304,162,000
1,601,304,162,000
CONTRIBUTOR
null
null
null
This happens when both 1. Huggingface datasets cache dir does not exist 2. Try to load a local dataset script builder.py throws an error when trying to create a filelock in a directory (cache/datasets) that does not exist https://github.com/huggingface/datasets/blob/master/src/datasets/builder.py#L177 Tested o...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/671/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/671/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/670
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/670/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/670/comments
https://api.github.com/repos/huggingface/datasets/issues/670/events
https://github.com/huggingface/datasets/pull/670
709,061,231
MDExOlB1bGxSZXF1ZXN0NDkzMTc4OTQw
670
Fix SQuAD metric kwargs description
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,601,050,137,000
1,601,395,059,000
1,601,395,058,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/670", "html_url": "https://github.com/huggingface/datasets/pull/670", "diff_url": "https://github.com/huggingface/datasets/pull/670.diff", "patch_url": "https://github.com/huggingface/datasets/pull/670.patch", "merged_at": 1601395057000 }
The `answer_start` field was missing in the kwargs docstring. This should fix #657 FYI another fix was proposed by @tshrjn in #658 and suggests to remove this field. However IMO `answer_start` is useful to match the squad dataset format for consistency, even though it is not used in the metric computation. I th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/670/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/670/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/669
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/669/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/669/comments
https://api.github.com/repos/huggingface/datasets/issues/669/events
https://github.com/huggingface/datasets/issues/669
708,857,595
MDU6SXNzdWU3MDg4NTc1OTU=
669
How to skip a example when running dataset.map
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi @xixiaoyao,\r\nDepending on what you want to do you can:\r\n- use a first step of `filter` to filter out the invalid examples: https://huggingface.co/docs/datasets/processing.html#filtering-rows-select-and-filter\r\n- or directly detect the invalid examples inside the callable used with `map` and return them un...
1,601,032,673,000
1,655,502,303,000
1,601,915,293,000
NONE
null
null
null
in processing func, I process examples and detect some invalid examples, which I did not want it to be added into train dataset. However I did not find how to skip this recognized invalid example when doing dataset.map.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/669/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/669/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/668
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/668/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/668/comments
https://api.github.com/repos/huggingface/datasets/issues/668/events
https://github.com/huggingface/datasets/issues/668
708,310,956
MDU6SXNzdWU3MDgzMTA5NTY=
668
OverflowError when slicing with an array containing negative ids
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,964,834,000
1,601,304,139,000
1,601,304,139,000
MEMBER
null
null
null
```python from datasets import Dataset d = ds.Dataset.from_dict({"a": range(10)}) print(d[0]) # {'a': 0} print(d[-1]) # {'a': 9} print(d[[0, -1]]) # OverflowError ``` results in ``` --------------------------------------------------------------------------- OverflowError ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/668/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/668/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/667
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/667/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/667/comments
https://api.github.com/repos/huggingface/datasets/issues/667/events
https://github.com/huggingface/datasets/issues/667
708,258,392
MDU6SXNzdWU3MDgyNTgzOTI=
667
Loss not decrease with Datasets and Transformers
{ "login": "wangcongcong123", "id": 23032865, "node_id": "MDQ6VXNlcjIzMDMyODY1", "avatar_url": "https://avatars.githubusercontent.com/u/23032865?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wangcongcong123", "html_url": "https://github.com/wangcongcong123", "followers_url": "https://api...
[]
closed
false
null
[]
null
[ "And I tested it on T5ForConditionalGeneration, that works no problem.", "Hi did you manage to fix your issue ?\r\n\r\nIf so feel free to share your fix and close this thread" ]
1,600,960,483,000
1,609,531,285,000
1,609,531,285,000
NONE
null
null
null
HI, The following script is used to fine-tune a BertForSequenceClassification model on SST2. The script is adapted from [this colab](https://colab.research.google.com/github/huggingface/datasets/blob/master/notebooks/Overview.ipynb) that presents an example of fine-tuning BertForQuestionAnswering using squad data...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/667/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/667/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/666
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/666/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/666/comments
https://api.github.com/repos/huggingface/datasets/issues/666/events
https://github.com/huggingface/datasets/issues/666
707,608,578
MDU6SXNzdWU3MDc2MDg1Nzg=
666
Does both 'bookcorpus' and 'wikipedia' belong to the same datasets which Google used for pretraining BERT?
{ "login": "wahab4114", "id": 31090427, "node_id": "MDQ6VXNlcjMxMDkwNDI3", "avatar_url": "https://avatars.githubusercontent.com/u/31090427?v=4", "gravatar_id": "", "url": "https://api.github.com/users/wahab4114", "html_url": "https://github.com/wahab4114", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "No they are other similar copies but they are not provided by the official Bert models authors." ]
1,600,887,745,000
1,603,811,965,000
1,603,811,965,000
NONE
null
null
null
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/666/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/666/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/665
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/665/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/665/comments
https://api.github.com/repos/huggingface/datasets/issues/665/events
https://github.com/huggingface/datasets/issues/665
707,037,738
MDU6SXNzdWU3MDcwMzc3Mzg=
665
runing dataset.map, it raises TypeError: can't pickle Tokenizer objects
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi !\r\nIt works on my side with both the LongFormerTokenizer and the LongFormerTokenizerFast.\r\n\r\nWhich version of transformers/datasets are you using ?", "transformers and datasets are both the latest", "Then I guess you need to give us more informations on your setup (OS, python, GPU, etc) or a Google Co...
1,600,835,294,000
1,602,149,536,000
1,602,149,536,000
NONE
null
null
null
I load squad dataset. Then want to process data use following function with `Huggingface Transformers LongformerTokenizer`. ``` def convert_to_features(example): # Tokenize contexts and questions (as pairs of inputs) input_pairs = [example['question'], example['context']] encodings = tokenizer.encode...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/665/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/665/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/664
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/664/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/664/comments
https://api.github.com/repos/huggingface/datasets/issues/664/events
https://github.com/huggingface/datasets/issues/664
707,017,791
MDU6SXNzdWU3MDcwMTc3OTE=
664
load_dataset from local squad.py, raise error: TypeError: 'NoneType' object is not callable
{ "login": "xixiaoyao", "id": 24541791, "node_id": "MDQ6VXNlcjI0NTQxNzkx", "avatar_url": "https://avatars.githubusercontent.com/u/24541791?v=4", "gravatar_id": "", "url": "https://api.github.com/users/xixiaoyao", "html_url": "https://github.com/xixiaoyao", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi !\r\nThanks for reporting.\r\nIt looks like no object inherits from `datasets.GeneratorBasedBuilder` (or more generally from `datasets.DatasetBuilder`) in your script.\r\n\r\nCould you check that there exist at least one dataset builder class ?", "Hi @xixiaoyao did you manage to fix your issue ?", "No activ...
1,600,833,216,000
1,603,184,773,000
1,603,184,773,000
NONE
null
null
null
version: 1.0.2 ``` train_dataset = datasets.load_dataset('squad') ``` The above code can works. However, when I download the squad.py from your server, and saved as `my_squad.py` to local. I run followings raise errors. ``` train_dataset = datasets.load_dataset('./my_squad.py') ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/664/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/664/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/663
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/663/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/663/comments
https://api.github.com/repos/huggingface/datasets/issues/663/events
https://github.com/huggingface/datasets/pull/663
706,732,636
MDExOlB1bGxSZXF1ZXN0NDkxMjI3NzUz
663
Created dataset card snli.md
{ "login": "mcmillanmajora", "id": 26722925, "node_id": "MDQ6VXNlcjI2NzIyOTI1", "avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mcmillanmajora", "html_url": "https://github.com/mcmillanmajora", "followers_url": "https://api.gi...
[ { "id": 2067401494, "node_id": "MDU6TGFiZWwyMDY3NDAxNDk0", "url": "https://api.github.com/repos/huggingface/datasets/labels/Dataset%20discussion", "name": "Dataset discussion", "color": "72f99f", "default": false, "description": "Discussions on the datasets" } ]
closed
false
{ "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api.github.com/users/yje...
[ { "login": "yjernite", "id": 10469459, "node_id": "MDQ6VXNlcjEwNDY5NDU5", "avatar_url": "https://avatars.githubusercontent.com/u/10469459?v=4", "gravatar_id": "", "url": "https://api.github.com/users/yjernite", "html_url": "https://github.com/yjernite", "followers_url": "https://api....
null
[ "Adding a direct link to the rendered markdown:\r\nhttps://github.com/mcmillanmajora/datasets/blob/add_dataset_documentation/datasets/snli/README.md\r\n", "It would be amazing if we ended up with this much information on all of our datasets :) \r\n\r\nI don't think there's too much repetition, everything that is ...
1,600,813,777,000
1,602,608,720,000
1,602,534,412,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/663", "html_url": "https://github.com/huggingface/datasets/pull/663", "diff_url": "https://github.com/huggingface/datasets/pull/663.diff", "patch_url": "https://github.com/huggingface/datasets/pull/663.patch", "merged_at": 1602534412000 }
First draft of a dataset card using the SNLI corpus as an example. This is mostly based on the [Google Doc draft](https://docs.google.com/document/d/1dKPGP-dA2W0QoTRGfqQ5eBp0CeSsTy7g2yM8RseHtos/edit), but I added a few sections and moved some things around. - I moved **Who Was Involved** to follow **Language**, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/663/reactions", "total_count": 3, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/663/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/662
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/662/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/662/comments
https://api.github.com/repos/huggingface/datasets/issues/662/events
https://github.com/huggingface/datasets/pull/662
706,689,866
MDExOlB1bGxSZXF1ZXN0NDkxMTkyNTM3
662
Created dataset card snli.md
{ "login": "mcmillanmajora", "id": 26722925, "node_id": "MDQ6VXNlcjI2NzIyOTI1", "avatar_url": "https://avatars.githubusercontent.com/u/26722925?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mcmillanmajora", "html_url": "https://github.com/mcmillanmajora", "followers_url": "https://api.gi...
[ { "id": 2067401494, "node_id": "MDU6TGFiZWwyMDY3NDAxNDk0", "url": "https://api.github.com/repos/huggingface/datasets/labels/Dataset%20discussion", "name": "Dataset discussion", "color": "72f99f", "default": false, "description": "Discussions on the datasets" } ]
closed
false
null
[]
null
[ "Resubmitting on a new fork" ]
1,600,808,417,000
1,600,809,981,000
1,600,809,981,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/662", "html_url": "https://github.com/huggingface/datasets/pull/662", "diff_url": "https://github.com/huggingface/datasets/pull/662.diff", "patch_url": "https://github.com/huggingface/datasets/pull/662.patch", "merged_at": null }
First draft of a dataset card using the SNLI corpus as an example
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/662/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/662/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/661
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/661/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/661/comments
https://api.github.com/repos/huggingface/datasets/issues/661/events
https://github.com/huggingface/datasets/pull/661
706,465,936
MDExOlB1bGxSZXF1ZXN0NDkxMDA3NjEw
661
Replace pa.OSFile by open
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,787,159,000
1,620,239,076,000
1,600,787,725,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/661", "html_url": "https://github.com/huggingface/datasets/pull/661", "diff_url": "https://github.com/huggingface/datasets/pull/661.diff", "patch_url": "https://github.com/huggingface/datasets/pull/661.patch", "merged_at": null }
It should fix #643
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/661/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/661/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/660
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/660/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/660/comments
https://api.github.com/repos/huggingface/datasets/issues/660/events
https://github.com/huggingface/datasets/pull/660
706,324,032
MDExOlB1bGxSZXF1ZXN0NDkwODkyMjQ0
660
add openwebtext
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "BTW, is there a one-line command to make our building scripts pass flake8 test? (included code quality test), I got like trailing space or mixed space and tab warning and error, and fixed them manually.", "> BTW, is there a one-line command to make our building scripts pass flake8 test? (included code quality te...
1,600,776,322,000
1,601,976,010,000
1,601,284,046,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/660", "html_url": "https://github.com/huggingface/datasets/pull/660", "diff_url": "https://github.com/huggingface/datasets/pull/660.diff", "patch_url": "https://github.com/huggingface/datasets/pull/660.patch", "merged_at": 1601284046000 }
This adds [The OpenWebText Corpus](https://skylion007.github.io/OpenWebTextCorpus/), which is a clean and large text corpus for nlp pretraining. It is an open source effort to reproduce OpenAI’s WebText dataset used by GPT-2, and it is also needed to reproduce ELECTRA. It solves #132 . ### Besides dataset buildin...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/660/reactions", "total_count": 2, "+1": 0, "-1": 0, "laugh": 0, "hooray": 1, "confused": 0, "heart": 0, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/660/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/659
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/659/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/659/comments
https://api.github.com/repos/huggingface/datasets/issues/659/events
https://github.com/huggingface/datasets/pull/659
706,231,506
MDExOlB1bGxSZXF1ZXN0NDkwODE4NTY1
659
Keep new columns in transmit format
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,768,043,000
1,600,769,242,000
1,600,769,240,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/659", "html_url": "https://github.com/huggingface/datasets/pull/659", "diff_url": "https://github.com/huggingface/datasets/pull/659.diff", "patch_url": "https://github.com/huggingface/datasets/pull/659.patch", "merged_at": 1600769240000 }
When a dataset is formatted with a list of columns that `__getitem__` should return, then calling `map` to add new columns doesn't add the new columns to this list. It caused `KeyError` issues in #620 I changed the logic to add those new columns to the list that `__getitem__` should return.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/659/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/659/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/658
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/658/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/658/comments
https://api.github.com/repos/huggingface/datasets/issues/658/events
https://github.com/huggingface/datasets/pull/658
706,206,247
MDExOlB1bGxSZXF1ZXN0NDkwNzk4MDc0
658
Fix squad metric's Features
{ "login": "tshrjn", "id": 8372098, "node_id": "MDQ6VXNlcjgzNzIwOTg=", "avatar_url": "https://avatars.githubusercontent.com/u/8372098?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tshrjn", "html_url": "https://github.com/tshrjn", "followers_url": "https://api.github.com/users/tshrjn/foll...
[]
closed
false
null
[]
null
[ "Closing this one in favor of #670 \r\n\r\nThanks again for reporting the issue and proposing this fix !\r\nLet me know if you have other remarks" ]
1,600,765,792,000
1,601,395,110,000
1,601,395,110,000
NONE
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/658", "html_url": "https://github.com/huggingface/datasets/pull/658", "diff_url": "https://github.com/huggingface/datasets/pull/658.diff", "patch_url": "https://github.com/huggingface/datasets/pull/658.patch", "merged_at": null }
Resolves issue [657](https://github.com/huggingface/datasets/issues/657).
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/658/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/658/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/657
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/657/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/657/comments
https://api.github.com/repos/huggingface/datasets/issues/657/events
https://github.com/huggingface/datasets/issues/657
706,204,383
MDU6SXNzdWU3MDYyMDQzODM=
657
Squad Metric Description & Feature Mismatch
{ "login": "tshrjn", "id": 8372098, "node_id": "MDQ6VXNlcjgzNzIwOTg=", "avatar_url": "https://avatars.githubusercontent.com/u/8372098?v=4", "gravatar_id": "", "url": "https://api.github.com/users/tshrjn", "html_url": "https://github.com/tshrjn", "followers_url": "https://api.github.com/users/tshrjn/foll...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nThere indeed a mismatch between the features and the kwargs description\r\n\r\nI believe `answer_start` was added to match the squad dataset format for consistency, even though it is not used in the metric computation. I think I'd rather keep it this way, so that you can just give `refere...
1,600,765,620,000
1,602,555,416,000
1,601,395,058,000
NONE
null
null
null
The [description](https://github.com/huggingface/datasets/blob/master/metrics/squad/squad.py#L39) doesn't mention `answer_start` in squad. However the `datasets.features` require [it](https://github.com/huggingface/datasets/blob/master/metrics/squad/squad.py#L68). It's also not used in the evaluation.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/657/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/657/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/656
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/656/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/656/comments
https://api.github.com/repos/huggingface/datasets/issues/656/events
https://github.com/huggingface/datasets/pull/656
705,736,319
MDExOlB1bGxSZXF1ZXN0NDkwNDEwODAz
656
Use multiprocess from pathos for multiprocessing
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "We can just install multiprocess actually, I'll change that", "Just an FYI: I remember that I wanted to try pathos a couple of years back and I ran into issues considering cross-platform; the code would just break on Windows. If I can verify this PR by running CPU tests on Windows, let me know!", "That's good ...
1,600,704,739,000
1,601,304,340,000
1,601,304,339,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/656", "html_url": "https://github.com/huggingface/datasets/pull/656", "diff_url": "https://github.com/huggingface/datasets/pull/656.diff", "patch_url": "https://github.com/huggingface/datasets/pull/656.patch", "merged_at": 1601304339000 }
[Multiprocess](https://github.com/uqfoundation/multiprocess) (from the [pathos](https://github.com/uqfoundation/pathos) project) allows to use lambda functions in multiprocessed map. It was suggested to use it by @kandorm. We're already using dill which is its only dependency.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/656/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/656/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/655
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/655/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/655/comments
https://api.github.com/repos/huggingface/datasets/issues/655/events
https://github.com/huggingface/datasets/pull/655
705,672,208
MDExOlB1bGxSZXF1ZXN0NDkwMzU4OTQ3
655
added Winogrande debiased subset
{ "login": "TevenLeScao", "id": 26709476, "node_id": "MDQ6VXNlcjI2NzA5NDc2", "avatar_url": "https://avatars.githubusercontent.com/u/26709476?v=4", "gravatar_id": "", "url": "https://api.github.com/users/TevenLeScao", "html_url": "https://github.com/TevenLeScao", "followers_url": "https://api.github.com/...
[]
closed
false
null
[]
null
[ "To fix the CI you just have to copy the dummy data to the 1.1.0 folder, and maybe create the dummy ones for the `debiased` configuration", "Fixed! Thanks @lhoestq " ]
1,600,699,868,000
1,600,705,240,000
1,600,704,964,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/655", "html_url": "https://github.com/huggingface/datasets/pull/655", "diff_url": "https://github.com/huggingface/datasets/pull/655.diff", "patch_url": "https://github.com/huggingface/datasets/pull/655.patch", "merged_at": 1600704964000 }
The [Winogrande](https://arxiv.org/abs/1907.10641) paper mentions a `debiased` subset that wasn't in the first release; this PR adds it.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/655/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/655/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/654
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/654/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/654/comments
https://api.github.com/repos/huggingface/datasets/issues/654/events
https://github.com/huggingface/datasets/pull/654
705,511,058
MDExOlB1bGxSZXF1ZXN0NDkwMjI1Nzk3
654
Allow empty inputs in metrics
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,687,596,000
1,601,956,308,000
1,600,704,818,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/654", "html_url": "https://github.com/huggingface/datasets/pull/654", "diff_url": "https://github.com/huggingface/datasets/pull/654.diff", "patch_url": "https://github.com/huggingface/datasets/pull/654.patch", "merged_at": 1600704818000 }
There was an arrow error when trying to compute a metric with empty inputs. The error was occurring when reading the arrow file, before calling metric._compute.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/654/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/654/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/653
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/653/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/653/comments
https://api.github.com/repos/huggingface/datasets/issues/653/events
https://github.com/huggingface/datasets/pull/653
705,482,391
MDExOlB1bGxSZXF1ZXN0NDkwMjAxOTg4
653
handle data alteration when trying type
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,684,909,000
1,600,704,786,000
1,600,704,785,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/653", "html_url": "https://github.com/huggingface/datasets/pull/653", "diff_url": "https://github.com/huggingface/datasets/pull/653.diff", "patch_url": "https://github.com/huggingface/datasets/pull/653.patch", "merged_at": 1600704785000 }
Fix #649 The bug came from the type inference that didn't handle a weird case in Pyarrow. Indeed this code runs without error but alters the data in arrow: ```python import pyarrow as pa type = pa.struct({"a": pa.struct({"b": pa.string()})}) array_with_altered_data = pa.array([{"a": {"b": "foo", "c": "bar"}}...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/653/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/653/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/652
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/652/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/652/comments
https://api.github.com/repos/huggingface/datasets/issues/652/events
https://github.com/huggingface/datasets/pull/652
705,390,850
MDExOlB1bGxSZXF1ZXN0NDkwMTI3MjIx
652
handle connection error in download_prepared_from_hf_gcs
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,676,471,000
1,600,676,923,000
1,600,676,922,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/652", "html_url": "https://github.com/huggingface/datasets/pull/652", "diff_url": "https://github.com/huggingface/datasets/pull/652.diff", "patch_url": "https://github.com/huggingface/datasets/pull/652.patch", "merged_at": 1600676922000 }
Fix #647
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/652/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/652/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/651
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/651/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/651/comments
https://api.github.com/repos/huggingface/datasets/issues/651/events
https://github.com/huggingface/datasets/issues/651
705,212,034
MDU6SXNzdWU3MDUyMTIwMzQ=
651
Problem with JSON dataset format
{ "login": "vikigenius", "id": 12724810, "node_id": "MDQ6VXNlcjEyNzI0ODEw", "avatar_url": "https://avatars.githubusercontent.com/u/12724810?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vikigenius", "html_url": "https://github.com/vikigenius", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[ "Currently the `json` dataset doesn't support this format unfortunately.\r\nHowever you could load it with\r\n```python\r\nfrom datasets import Dataset\r\nimport pandas as pd\r\n\r\ndf = pd.read_json(\"path_to_local.json\", orient=\"index\")\r\ndataset = Dataset.from_pandas(df)\r\n```", "or you can make a custom ...
1,600,646,234,000
1,600,690,464,000
null
NONE
null
null
null
I have a local json dataset with the following form. { 'id01234': {'key1': value1, 'key2': value2, 'key3': value3}, 'id01235': {'key1': value1, 'key2': value2, 'key3': value3}, . . . 'id09999': {'key1': value1, 'key2': value2, 'key3': value3} } Note that instead of a list of records i...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/651/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/651/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/650
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/650/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/650/comments
https://api.github.com/repos/huggingface/datasets/issues/650/events
https://github.com/huggingface/datasets/issues/650
704,861,844
MDU6SXNzdWU3MDQ4NjE4NDQ=
650
dummy data testing can't test datasets using `dl_manager.extract` in `_split_generators`
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Hi :) \r\nIn your dummy data zip file you can just have `subset000.xz` as directories instead of compressed files.\r\nLet me know if it helps", "Thanks for your comment @lhoestq ,\r\nJust for confirmation, changing dummy data like this won't make dummy test test the functionality to extract `subsetxxx.xz` but ac...
1,600,513,623,000
1,600,775,650,000
1,600,775,649,000
CONTRIBUTOR
null
null
null
Hi, I recently want to add a dataset whose source data is like this ``` openwebtext.tar.xz |__ openwebtext |__subset000.xz | |__ ....txt | |__ ....txt | ... |__ subset001.xz | .... ``` So I wrote `openwebtext.py` like this ``` d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/650/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/650/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/649
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/649/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/649/comments
https://api.github.com/repos/huggingface/datasets/issues/649/events
https://github.com/huggingface/datasets/issues/649
704,838,415
MDU6SXNzdWU3MDQ4Mzg0MTU=
649
Inconsistent behavior in map
{ "login": "krandiash", "id": 10166085, "node_id": "MDQ6VXNlcjEwMTY2MDg1", "avatar_url": "https://avatars.githubusercontent.com/u/10166085?v=4", "gravatar_id": "", "url": "https://api.github.com/users/krandiash", "html_url": "https://github.com/krandiash", "followers_url": "https://api.github.com/users/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Thanks for reporting !\r\n\r\nThis issue must have appeared when we refactored type inference in `nlp`\r\nBy default the library tries to keep the same feature types when applying `map` but apparently it has troubles with nested structures. I'll try to fix that next week" ]
1,600,504,872,000
1,600,704,785,000
1,600,704,785,000
NONE
null
null
null
I'm observing inconsistent behavior when applying .map(). This happens specifically when I'm incrementally adding onto a feature that is a nested dictionary. Here's a simple example that reproduces the problem. ```python import datasets # Dataset with a single feature called 'field' consisting of two examples d...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/649/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/649/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/648
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/648/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/648/comments
https://api.github.com/repos/huggingface/datasets/issues/648/events
https://github.com/huggingface/datasets/issues/648
704,753,123
MDU6SXNzdWU3MDQ3NTMxMjM=
648
offset overflow when multiprocessing batched map on large datasets.
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "This should be fixed with #645 ", "Feel free to re-open if it still occurs" ]
1,600,481,711,000
1,600,534,027,000
1,600,533,991,000
CONTRIBUTOR
null
null
null
It only happened when "multiprocessing" + "batched" + "large dataset" at the same time. ``` def bprocess(examples): examples['len'] = [] for text in examples['text']: examples['len'].append(len(text)) return examples wiki.map(brpocess, batched=True, num_proc=8) ``` ``` ----------------------------...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/648/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/648/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/647
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/647/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/647/comments
https://api.github.com/repos/huggingface/datasets/issues/647/events
https://github.com/huggingface/datasets/issues/647
704,734,764
MDU6SXNzdWU3MDQ3MzQ3NjQ=
647
Cannot download dataset_info.json
{ "login": "chiyuzhang94", "id": 33407613, "node_id": "MDQ6VXNlcjMzNDA3NjEz", "avatar_url": "https://avatars.githubusercontent.com/u/33407613?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chiyuzhang94", "html_url": "https://github.com/chiyuzhang94", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWe should add support for servers without internet connection indeed\r\nI'll do that early next week", "Thanks, @lhoestq !\r\nPlease let me know when it is available. ", "Right now the recommended way is to create the dataset on a server with internet connection and then to save it an...
1,600,479,315,000
1,600,676,922,000
1,600,676,922,000
NONE
null
null
null
I am running my job on a cloud server where does not provide for connections from the standard compute nodes to outside resources. Hence, when I use `dataset.load_dataset()` to load data, I got an error like this: ``` ConnectionError: Couldn't reach https://storage.googleapis.com/huggingface-nlp/cache/datasets/text...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/647/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/647/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/646
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/646/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/646/comments
https://api.github.com/repos/huggingface/datasets/issues/646/events
https://github.com/huggingface/datasets/pull/646
704,607,371
MDExOlB1bGxSZXF1ZXN0NDg5NTAyMTM3
646
Fix docs typos
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[]
1,600,457,547,000
1,600,705,854,000
1,600,704,852,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/646", "html_url": "https://github.com/huggingface/datasets/pull/646", "diff_url": "https://github.com/huggingface/datasets/pull/646.diff", "patch_url": "https://github.com/huggingface/datasets/pull/646.patch", "merged_at": 1600704852000 }
This PR fixes few typos in the docs and the error in the code snippet in the set_format section in docs/source/torch_tensorflow.rst. `torch.utils.data.Dataloader` expects padded batches so it throws an error due to not being able to stack the unpadded tensors. If we follow the Quick tour from the docs where they add th...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/646/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/646/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/645
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/645/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/645/comments
https://api.github.com/repos/huggingface/datasets/issues/645/events
https://github.com/huggingface/datasets/pull/645
704,542,234
MDExOlB1bGxSZXF1ZXN0NDg5NDQ5MjAx
645
Don't use take on dataset table in pyarrow 1.0.x
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I tried lower batch sizes and it didn't accelerate filter (quite the opposite actually).\r\nThe slow-down also appears for pyarrow 0.17.1 for some reason, not sure it comes from these changes", "I just checked the benchmarks of other PRs and some of them had 300s (!!) for filter. This needs some investigation.."...
1,600,450,294,000
1,600,533,992,000
1,600,533,991,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/645", "html_url": "https://github.com/huggingface/datasets/pull/645", "diff_url": "https://github.com/huggingface/datasets/pull/645.diff", "patch_url": "https://github.com/huggingface/datasets/pull/645.patch", "merged_at": 1600533991000 }
Fix #615
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/645/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/645/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/644
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/644/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/644/comments
https://api.github.com/repos/huggingface/datasets/issues/644/events
https://github.com/huggingface/datasets/pull/644
704,534,501
MDExOlB1bGxSZXF1ZXN0NDg5NDQzMTk1
644
Better windows support
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "This PR is ready :)\r\nIt brings official support for windows.\r\n\r\nSome tests `AWSDatasetTest` are failing.\r\nThis is because I had to fix a few datasets that were not compatible with windows.\r\nThese test will pass once they got merged on master :)" ]
1,600,449,456,000
1,601,042,550,000
1,601,042,548,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/644", "html_url": "https://github.com/huggingface/datasets/pull/644", "diff_url": "https://github.com/huggingface/datasets/pull/644.diff", "patch_url": "https://github.com/huggingface/datasets/pull/644.patch", "merged_at": 1601042548000 }
There are a few differences in the behavior of python and pyarrow on windows. For example there are restrictions when accessing/deleting files that are open Fix #590
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/644/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/644/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/643
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/643/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/643/comments
https://api.github.com/repos/huggingface/datasets/issues/643/events
https://github.com/huggingface/datasets/issues/643
704,477,164
MDU6SXNzdWU3MDQ0NzcxNjQ=
643
Caching processed dataset at wrong folder
{ "login": "mrm8488", "id": 3653789, "node_id": "MDQ6VXNlcjM2NTM3ODk=", "avatar_url": "https://avatars.githubusercontent.com/u/3653789?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mrm8488", "html_url": "https://github.com/mrm8488", "followers_url": "https://api.github.com/users/mrm8488/...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nIt uses a temporary file to write the data.\r\nHowever it looks like the temporary file is not placed in the right directory during the processing", "Well actually I just tested and the temporary file is placed in the same directory, so it should work as expected.\r\nWhich version of `d...
1,600,443,686,000
1,645,023,209,000
1,645,023,209,000
CONTRIBUTOR
null
null
null
Hi guys, I run this on my Colab (PRO): ```python from datasets import load_dataset dataset = load_dataset('text', data_files='/content/corpus.txt', cache_dir='/content/drive/My Drive', split='train') def encode(examples): return tokenizer(examples['text'], truncation=True, padding='max_length') dataset = ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/643/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/643/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/642
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/642/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/642/comments
https://api.github.com/repos/huggingface/datasets/issues/642/events
https://github.com/huggingface/datasets/pull/642
704,397,499
MDExOlB1bGxSZXF1ZXN0NDg5MzMwMDAx
642
Rename wnut fields
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,437,091,000
1,600,449,511,000
1,600,449,510,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/642", "html_url": "https://github.com/huggingface/datasets/pull/642", "diff_url": "https://github.com/huggingface/datasets/pull/642.diff", "patch_url": "https://github.com/huggingface/datasets/pull/642.patch", "merged_at": 1600449510000 }
As mentioned in #641 it would be cool to have it follow the naming of the other NER datasets
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/642/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/642/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/641
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/641/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/641/comments
https://api.github.com/repos/huggingface/datasets/issues/641/events
https://github.com/huggingface/datasets/pull/641
704,373,940
MDExOlB1bGxSZXF1ZXN0NDg5MzExOTU3
641
Add Polyglot-NER Dataset
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/...
[]
closed
false
null
[]
null
[ "Hi @joeddav thanks for adding this! (I did a long webarchive.org session to actually find that dataset a while ago).\r\n\r\nOne question: should we manually correct the labeling scheme to (at least) IOB1?\r\n\r\nThat means \"LOC\" will be converted to \"I-LOC\". IOB1 is not explict. mentioned in the paper, but it ...
1,600,435,304,000
1,600,571,083,000
1,600,571,083,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/641", "html_url": "https://github.com/huggingface/datasets/pull/641", "diff_url": "https://github.com/huggingface/datasets/pull/641.diff", "patch_url": "https://github.com/huggingface/datasets/pull/641.patch", "merged_at": 1600571083000 }
Adds the [Polyglot-NER dataset](https://sites.google.com/site/rmyeid/projects/polylgot-ner) with named entity tags for 40 languages. I include separate configs for each language as well as a `combined` config which lumps them all together.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/641/reactions", "total_count": 6, "+1": 2, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 2, "rocket": 2, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/641/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/640
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/640/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/640/comments
https://api.github.com/repos/huggingface/datasets/issues/640/events
https://github.com/huggingface/datasets/pull/640
704,311,758
MDExOlB1bGxSZXF1ZXN0NDg5MjYwNTc1
640
Make shuffle compatible with temp_seed
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,429,138,000
1,600,429,671,000
1,600,429,670,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/640", "html_url": "https://github.com/huggingface/datasets/pull/640", "diff_url": "https://github.com/huggingface/datasets/pull/640.diff", "patch_url": "https://github.com/huggingface/datasets/pull/640.patch", "merged_at": 1600429670000 }
This code used to return different dataset at each run ```python import dataset as ds dataset = ... with ds.temp_seed(42): shuffled = dataset.shuffle() ``` Now it returns the same one since the seed is set
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/640/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/640/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/639
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/639/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/639/comments
https://api.github.com/repos/huggingface/datasets/issues/639/events
https://github.com/huggingface/datasets/pull/639
704,217,963
MDExOlB1bGxSZXF1ZXN0NDg5MTgxOTY3
639
Update glue QQP checksum
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,420,095,000
1,600,429,028,000
1,600,429,027,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/639", "html_url": "https://github.com/huggingface/datasets/pull/639", "diff_url": "https://github.com/huggingface/datasets/pull/639.diff", "patch_url": "https://github.com/huggingface/datasets/pull/639.patch", "merged_at": 1600429027000 }
Fix #638
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/639/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/639/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/638
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/638/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/638/comments
https://api.github.com/repos/huggingface/datasets/issues/638/events
https://github.com/huggingface/datasets/issues/638
704,146,956
MDU6SXNzdWU3MDQxNDY5NTY=
638
GLUE/QQP dataset: NonMatchingChecksumError
{ "login": "richarddwang", "id": 17963619, "node_id": "MDQ6VXNlcjE3OTYzNjE5", "avatar_url": "https://avatars.githubusercontent.com/u/17963619?v=4", "gravatar_id": "", "url": "https://api.github.com/users/richarddwang", "html_url": "https://github.com/richarddwang", "followers_url": "https://api.github.c...
[]
closed
false
null
[]
null
[ "Hi ! Sure I'll take a look" ]
1,600,412,950,000
1,600,429,027,000
1,600,429,027,000
CONTRIBUTOR
null
null
null
Hi @lhoestq , I know you are busy and there are also other important issues. But if this is easy to be fixed, I am shamelessly wondering if you can give me some help , so I can evaluate my models and restart with my developing cycle asap. 😚 datasets version: editable install of master at 9/17 `datasets.load_data...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/638/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/638/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/637
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/637/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/637/comments
https://api.github.com/repos/huggingface/datasets/issues/637/events
https://github.com/huggingface/datasets/pull/637
703,539,909
MDExOlB1bGxSZXF1ZXN0NDg4NjMwNzk4
637
Add MATINF
{ "login": "JetRunner", "id": 22514219, "node_id": "MDQ6VXNlcjIyNTE0MjE5", "avatar_url": "https://avatars.githubusercontent.com/u/22514219?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JetRunner", "html_url": "https://github.com/JetRunner", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[]
1,600,345,493,000
1,600,348,998,000
1,600,348,997,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/637", "html_url": "https://github.com/huggingface/datasets/pull/637", "diff_url": "https://github.com/huggingface/datasets/pull/637.diff", "patch_url": "https://github.com/huggingface/datasets/pull/637.patch", "merged_at": 1600348997000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/637/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/637/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/636
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/636/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/636/comments
https://api.github.com/repos/huggingface/datasets/issues/636/events
https://github.com/huggingface/datasets/pull/636
702,883,989
MDExOlB1bGxSZXF1ZXN0NDg4MDg3OTA5
636
Consistent ner features
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,600,271,785,000
1,600,336,379,000
1,600,336,378,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/636", "html_url": "https://github.com/huggingface/datasets/pull/636", "diff_url": "https://github.com/huggingface/datasets/pull/636.diff", "patch_url": "https://github.com/huggingface/datasets/pull/636.patch", "merged_at": 1600336378000 }
As discussed in #613 , this PR aims at making NER feature names consistent across datasets. I changed the feature names of LinCE and XTREME/PAN-X
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/636/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/636/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/635
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/635/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/635/comments
https://api.github.com/repos/huggingface/datasets/issues/635/events
https://github.com/huggingface/datasets/pull/635
702,822,439
MDExOlB1bGxSZXF1ZXN0NDg4MDM2OTE5
635
Loglevel
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I think it's ready now @stas00, did you want to add something else ?\r\nThis PR includes your changes but with the level set to warning", "LGTM, thank you, @lhoestq " ]
1,600,267,073,000
1,600,336,339,000
1,600,336,338,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/635", "html_url": "https://github.com/huggingface/datasets/pull/635", "diff_url": "https://github.com/huggingface/datasets/pull/635.diff", "patch_url": "https://github.com/huggingface/datasets/pull/635.patch", "merged_at": 1600336338000 }
Continuation of #618
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/635/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/635/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/634
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/634/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/634/comments
https://api.github.com/repos/huggingface/datasets/issues/634/events
https://github.com/huggingface/datasets/pull/634
702,676,041
MDExOlB1bGxSZXF1ZXN0NDg3OTEzOTk4
634
Add ConLL-2000 dataset
{ "login": "vblagoje", "id": 458335, "node_id": "MDQ6VXNlcjQ1ODMzNQ==", "avatar_url": "https://avatars.githubusercontent.com/u/458335?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vblagoje", "html_url": "https://github.com/vblagoje", "followers_url": "https://api.github.com/users/vblagoj...
[]
closed
false
null
[]
null
[]
1,600,254,851,000
1,600,339,090,000
1,600,339,090,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/634", "html_url": "https://github.com/huggingface/datasets/pull/634", "diff_url": "https://github.com/huggingface/datasets/pull/634.diff", "patch_url": "https://github.com/huggingface/datasets/pull/634.patch", "merged_at": 1600339090000 }
Adds ConLL-2000 dataset used for text chunking. See https://www.clips.uantwerpen.be/conll2000/chunking/ for details and [motivation](https://github.com/huggingface/transformers/pull/7041#issuecomment-692710948) behind this PR
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/634/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/634/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/633
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/633/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/633/comments
https://api.github.com/repos/huggingface/datasets/issues/633/events
https://github.com/huggingface/datasets/issues/633
702,440,484
MDU6SXNzdWU3MDI0NDA0ODQ=
633
Load large text file for LM pre-training resulting in OOM
{ "login": "leethu2012", "id": 29704017, "node_id": "MDQ6VXNlcjI5NzA0MDE3", "avatar_url": "https://avatars.githubusercontent.com/u/29704017?v=4", "gravatar_id": "", "url": "https://api.github.com/users/leethu2012", "html_url": "https://github.com/leethu2012", "followers_url": "https://api.github.com/use...
[]
open
false
null
[]
null
[ "Not sure what could cause that on the `datasets` side. Could this be a `Trainer` issue ? cc @julien-c @sgugger ?", "There was a memory leak issue fixed recently in master. You should install from source and see if it fixes your problem.", "@lhoestq @sgugger Thanks for your comments. I have install from source ...
1,600,230,795,000
1,613,476,921,000
null
NONE
null
null
null
I tried to pretrain Longformer using transformers and datasets. But I got OOM issues with loading a large text file. My script is almost like this: ```python from datasets import load_dataset @dataclass class DataCollatorForDatasetsLanguageModeling(DataCollatorForLanguageModeling): """ Data collator u...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/633/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/633/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/632
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/632/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/632/comments
https://api.github.com/repos/huggingface/datasets/issues/632/events
https://github.com/huggingface/datasets/pull/632
702,358,124
MDExOlB1bGxSZXF1ZXN0NDg3NjQ5OTQ2
632
Fix typos in the loading datasets docs
{ "login": "mariosasko", "id": 47462742, "node_id": "MDQ6VXNlcjQ3NDYyNzQy", "avatar_url": "https://avatars.githubusercontent.com/u/47462742?v=4", "gravatar_id": "", "url": "https://api.github.com/users/mariosasko", "html_url": "https://github.com/mariosasko", "followers_url": "https://api.github.com/use...
[]
closed
false
null
[]
null
[ "thanks!" ]
1,600,216,061,000
1,600,705,871,000
1,600,239,164,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/632", "html_url": "https://github.com/huggingface/datasets/pull/632", "diff_url": "https://github.com/huggingface/datasets/pull/632.diff", "patch_url": "https://github.com/huggingface/datasets/pull/632.patch", "merged_at": 1600239164000 }
This PR fixes two typos in the loading datasets docs, one of them being a broken link to the `load_dataset` function.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/632/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/632/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/631
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/631/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/631/comments
https://api.github.com/repos/huggingface/datasets/issues/631/events
https://github.com/huggingface/datasets/pull/631
701,711,255
MDExOlB1bGxSZXF1ZXN0NDg3MTE3OTA0
631
Fix text delimiter
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Which OS are you using ?@abhi1nandy2", "> Which OS are you using ?\r\n\r\nPRETTY_NAME=\"Debian GNU/Linux 9 (stretch)\"\r\nNAME=\"Debian GNU/Linux\"\r\nVERSION_ID=\"9\"\r\nVERSION=\"9 (stretch)\"\r\nVERSION_CODENAME=stretch\r\nID=debian\r\nHOME_URL=\"https://www.debian.org/\"\r\nSUPPORT_URL=\"https://www.debian.o...
1,600,157,322,000
1,600,786,986,000
1,600,158,385,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/631", "html_url": "https://github.com/huggingface/datasets/pull/631", "diff_url": "https://github.com/huggingface/datasets/pull/631.diff", "patch_url": "https://github.com/huggingface/datasets/pull/631.patch", "merged_at": 1600158385000 }
I changed the delimiter in the `text` dataset script. It should fix the `pyarrow.lib.ArrowInvalid: CSV parse error` from #622 I changed the delimiter to an unused ascii character that is not present in text files : `\b`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/631/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/631/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/630
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/630/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/630/comments
https://api.github.com/repos/huggingface/datasets/issues/630/events
https://github.com/huggingface/datasets/issues/630
701,636,350
MDU6SXNzdWU3MDE2MzYzNTA=
630
Text dataset not working with large files
{ "login": "ksjae", "id": 17930170, "node_id": "MDQ6VXNlcjE3OTMwMTcw", "avatar_url": "https://avatars.githubusercontent.com/u/17930170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ksjae", "html_url": "https://github.com/ksjae", "followers_url": "https://api.github.com/users/ksjae/follow...
[]
closed
false
null
[]
null
[ "Seems like it works when setting ```block_size=2100000000``` or something arbitrarily large though.", "Can you give us some stats on the data files you use as inputs?", "Basically ~600MB txt files(UTF-8) * 59. \r\ncontents like ```안녕하세요, 이것은 예제로 한번 말해보는 텍스트입니다. 그냥 이렇다고요.<|endoftext|>\\n```\r\n\r\nAlso, it gets...
1,600,149,756,000
1,601,072,503,000
1,601,072,503,000
NONE
null
null
null
``` Traceback (most recent call last): File "examples/language-modeling/run_language_modeling.py", line 333, in <module> main() File "examples/language-modeling/run_language_modeling.py", line 262, in main get_dataset(data_args, tokenizer=tokenizer, cache_dir=model_args.cache_dir) if training_args.do_t...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/630/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/630/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/629
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/629/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/629/comments
https://api.github.com/repos/huggingface/datasets/issues/629/events
https://github.com/huggingface/datasets/issues/629
701,517,550
MDU6SXNzdWU3MDE1MTc1NTA=
629
straddling object straddles two block boundaries
{ "login": "bharaniabhishek123", "id": 17970177, "node_id": "MDQ6VXNlcjE3OTcwMTc3", "avatar_url": "https://avatars.githubusercontent.com/u/17970177?v=4", "gravatar_id": "", "url": "https://api.github.com/users/bharaniabhishek123", "html_url": "https://github.com/bharaniabhishek123", "followers_url": "ht...
[]
closed
false
null
[]
null
[ "sorry it's an apache arrow issue." ]
1,600,129,846,000
1,600,130,177,000
1,600,129,937,000
NONE
null
null
null
I am trying to read json data (it's an array with lots of dictionaries) and getting block boundaries issue as below : I tried calling read_json with readOptions but no luck . ``` table = json.read_json(fn) Traceback (most recent call last): File "<stdin>", line 1, in <module> File "pyarrow/_json.pyx", li...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/629/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/629/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/628
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/628/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/628/comments
https://api.github.com/repos/huggingface/datasets/issues/628/events
https://github.com/huggingface/datasets/pull/628
701,496,053
MDExOlB1bGxSZXF1ZXN0NDg2OTQyNzgx
628
Update docs links in the contribution guideline
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/...
[]
closed
false
null
[]
null
[ "Thanks!" ]
1,600,126,039,000
1,604,351,003,000
1,600,150,775,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/628", "html_url": "https://github.com/huggingface/datasets/pull/628", "diff_url": "https://github.com/huggingface/datasets/pull/628.diff", "patch_url": "https://github.com/huggingface/datasets/pull/628.patch", "merged_at": 1600150775000 }
Fixed the `add a dataset` and `share a dataset` links in the contribution guideline to refer to the new docs website.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/628/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/628/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/627
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/627/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/627/comments
https://api.github.com/repos/huggingface/datasets/issues/627/events
https://github.com/huggingface/datasets/pull/627
701,411,661
MDExOlB1bGxSZXF1ZXN0NDg2ODcxMTg2
627
fix (#619) MLQA features names
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/...
[]
closed
false
null
[]
null
[]
1,600,116,119,000
1,604,351,072,000
1,600,239,251,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/627", "html_url": "https://github.com/huggingface/datasets/pull/627", "diff_url": "https://github.com/huggingface/datasets/pull/627.diff", "patch_url": "https://github.com/huggingface/datasets/pull/627.patch", "merged_at": 1600239251000 }
Fixed the features names as suggested in (#619) in the `_generate_examples` and `_info` methods in the MLQA loading script and also changed the names in the `dataset_infos.json` file.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/627/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/627/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/626
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/626/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/626/comments
https://api.github.com/repos/huggingface/datasets/issues/626/events
https://github.com/huggingface/datasets/pull/626
701,352,605
MDExOlB1bGxSZXF1ZXN0NDg2ODIzMTY1
626
Update GLUE URLs (now hosted on FB)
{ "login": "jeswan", "id": 57466294, "node_id": "MDQ6VXNlcjU3NDY2Mjk0", "avatar_url": "https://avatars.githubusercontent.com/u/57466294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jeswan", "html_url": "https://github.com/jeswan", "followers_url": "https://api.github.com/users/jeswan/fo...
[]
closed
false
null
[]
null
[]
1,600,110,339,000
1,600,239,198,000
1,600,239,198,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/626", "html_url": "https://github.com/huggingface/datasets/pull/626", "diff_url": "https://github.com/huggingface/datasets/pull/626.diff", "patch_url": "https://github.com/huggingface/datasets/pull/626.patch", "merged_at": 1600239198000 }
NYU is switching dataset hosting from Google to FB. This PR closes https://github.com/huggingface/datasets/issues/608 and is necessary for https://github.com/jiant-dev/jiant/issues/161. This PR updates the data URLs based on changes made in https://github.com/nyu-mll/jiant/pull/1112. Note: rebased on huggingface/dat...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/626/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/626/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/625
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/625/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/625/comments
https://api.github.com/repos/huggingface/datasets/issues/625/events
https://github.com/huggingface/datasets/issues/625
701,057,799
MDU6SXNzdWU3MDEwNTc3OTk=
625
dtype of tensors should be preserved
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users...
[]
closed
false
null
[]
null
[ "Indeed we convert tensors to list to be able to write in arrow format. Because of this conversion we lose the dtype information. We should add the dtype detection when we do type inference. However it would require a bit of refactoring since currently the conversion happens before the type inference..\r\n\r\nAnd t...
1,600,087,085,000
1,629,189,004,000
1,629,189,004,000
CONTRIBUTOR
null
null
null
After switching to `datasets` my model just broke. After a weekend of debugging, the issue was that my model could not handle the double that the Dataset provided, as it expected a float (but didn't give a warning, which seems a [PyTorch issue](https://discuss.pytorch.org/t/is-it-required-that-input-and-hidden-for-gru-...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/625/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/625/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/624
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/624/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/624/comments
https://api.github.com/repos/huggingface/datasets/issues/624/events
https://github.com/huggingface/datasets/issues/624
700,541,628
MDU6SXNzdWU3MDA1NDE2Mjg=
624
Add learningq dataset
{ "login": "krrishdholakia", "id": 17561003, "node_id": "MDQ6VXNlcjE3NTYxMDAz", "avatar_url": "https://avatars.githubusercontent.com/u/17561003?v=4", "gravatar_id": "", "url": "https://api.github.com/users/krrishdholakia", "html_url": "https://github.com/krrishdholakia", "followers_url": "https://api.gi...
[ { "id": 2067376369, "node_id": "MDU6TGFiZWwyMDY3Mzc2MzY5", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20request", "name": "dataset request", "color": "e99695", "default": false, "description": "Requesting to add a new dataset" } ]
open
false
null
[]
null
[]
1,599,992,427,000
1,600,077,002,000
null
NONE
null
null
null
Hi, Thank you again for this amazing repo. Would it be possible for y'all to add the LearningQ dataset - https://github.com/AngusGLChen/LearningQ ?
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/624/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/624/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/623
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/623/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/623/comments
https://api.github.com/repos/huggingface/datasets/issues/623/events
https://github.com/huggingface/datasets/issues/623
700,235,308
MDU6SXNzdWU3MDAyMzUzMDg=
623
Custom feature types in `load_dataset` from CSV
{ "login": "lvwerra", "id": 8264887, "node_id": "MDQ6VXNlcjgyNjQ4ODc=", "avatar_url": "https://avatars.githubusercontent.com/u/8264887?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lvwerra", "html_url": "https://github.com/lvwerra", "followers_url": "https://api.github.com/users/lvwerra/...
[ { "id": 1935892871, "node_id": "MDU6TGFiZWwxOTM1ODkyODcx", "url": "https://api.github.com/repos/huggingface/datasets/labels/enhancement", "name": "enhancement", "color": "a2eeef", "default": true, "description": "New feature or request" } ]
closed
false
null
[]
null
[ "Currently `csv` doesn't support the `features` attribute (unlike `json`).\r\nWhat you can do for now is cast the features using the in-place transform `cast_`\r\n\r\n```python\r\nfrom datasets import load_dataset\r\n\r\ndataset = load_dataset('csv', data_files=file_dict, delimiter=';', column_names=['text', 'label...
1,599,916,894,000
1,601,495,503,000
1,601,455,194,000
MEMBER
null
null
null
I am trying to load a local file with the `load_dataset` function and I want to predefine the feature types with the `features` argument. However, the types are always the same independent of the value of `features`. I am working with the local files from the emotion dataset. To get the data you can use the followi...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/623/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/623/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/622
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/622/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/622/comments
https://api.github.com/repos/huggingface/datasets/issues/622/events
https://github.com/huggingface/datasets/issues/622
700,225,826
MDU6SXNzdWU3MDAyMjU4MjY=
622
load_dataset for text files not working
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users...
[ { "id": 2067388877, "node_id": "MDU6TGFiZWwyMDY3Mzg4ODc3", "url": "https://api.github.com/repos/huggingface/datasets/labels/dataset%20bug", "name": "dataset bug", "color": "2edb81", "default": false, "description": "A bug in a dataset script provided in the library" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "Can you give us more information on your os and pip environments (pip list)?", "@thomwolf Sure. I'll try downgrading to 3.7 now even though Arrow say they support >=3.5.\r\n\r\nLinux (Ubuntu 18.04) - Python 3.8\r\n======================\r\nPackage - Version\r\n---------------------\r\ncertifi 2...
1,599,914,968,000
1,603,883,251,000
1,603,883,250,000
CONTRIBUTOR
null
null
null
Trying the following snippet, I get different problems on Linux and Windows. ```python dataset = load_dataset("text", data_files="data.txt") # or dataset = load_dataset("text", data_files=["data.txt"]) ``` (ps [This example](https://huggingface.co/docs/datasets/loading_datasets.html#json-files) shows that ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/622/reactions", "total_count": 3, "+1": 3, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/622/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/621
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/621/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/621/comments
https://api.github.com/repos/huggingface/datasets/issues/621/events
https://github.com/huggingface/datasets/pull/621
700,171,097
MDExOlB1bGxSZXF1ZXN0NDg1ODQ3ODYz
621
[docs] Index: The native emoji looks kinda ugly in large size
{ "login": "julien-c", "id": 326577, "node_id": "MDQ6VXNlcjMyNjU3Nw==", "avatar_url": "https://avatars.githubusercontent.com/u/326577?v=4", "gravatar_id": "", "url": "https://api.github.com/users/julien-c", "html_url": "https://github.com/julien-c", "followers_url": "https://api.github.com/users/julien-...
[]
closed
false
null
[]
null
[]
1,599,904,120,000
1,600,150,803,000
1,600,150,802,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/621", "html_url": "https://github.com/huggingface/datasets/pull/621", "diff_url": "https://github.com/huggingface/datasets/pull/621.diff", "patch_url": "https://github.com/huggingface/datasets/pull/621.patch", "merged_at": 1600150802000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/621/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/621/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/620
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/620/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/620/comments
https://api.github.com/repos/huggingface/datasets/issues/620/events
https://github.com/huggingface/datasets/issues/620
699,815,135
MDU6SXNzdWU2OTk4MTUxMzU=
620
map/filter multiprocessing raises errors and corrupts datasets
{ "login": "timothyjlaurent", "id": 2000204, "node_id": "MDQ6VXNlcjIwMDAyMDQ=", "avatar_url": "https://avatars.githubusercontent.com/u/2000204?v=4", "gravatar_id": "", "url": "https://api.github.com/users/timothyjlaurent", "html_url": "https://github.com/timothyjlaurent", "followers_url": "https://api.g...
[ { "id": 1935892857, "node_id": "MDU6TGFiZWwxOTM1ODkyODU3", "url": "https://api.github.com/repos/huggingface/datasets/labels/bug", "name": "bug", "color": "d73a4a", "default": true, "description": "Something isn't working" } ]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "It seems that I ran into the same problem\r\n```\r\ndef tokenize(cols, example):\r\n for in_col, out_col in cols.items():\r\n example[out_col] = hf_tokenizer.convert_tokens_to_ids(hf_tokenizer.tokenize(example[in_col]))\r\n return example\r\ncola = datasets.load_dataset('glue', 'cola')\r\ntokenized_cola = col...
1,599,863,406,000
1,602,174,707,000
1,602,174,706,000
NONE
null
null
null
After upgrading to the 1.0 started seeing errors in my data loading script after enabling multiprocessing. ```python ... ner_ds_dict = ner_ds.train_test_split(test_size=test_pct, shuffle=True, seed=seed) ner_ds_dict["validation"] = ner_ds_dict["test"] rel_ds_dict = rel_ds.train_test_split(test_si...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/620/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/620/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/619
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/619/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/619/comments
https://api.github.com/repos/huggingface/datasets/issues/619/events
https://github.com/huggingface/datasets/issues/619
699,733,612
MDU6SXNzdWU2OTk3MzM2MTI=
619
Mistakes in MLQA features names
{ "login": "M-Salti", "id": 9285264, "node_id": "MDQ6VXNlcjkyODUyNjQ=", "avatar_url": "https://avatars.githubusercontent.com/u/9285264?v=4", "gravatar_id": "", "url": "https://api.github.com/users/M-Salti", "html_url": "https://github.com/M-Salti", "followers_url": "https://api.github.com/users/M-Salti/...
[]
closed
false
null
[]
null
[ "Indeed you're right ! Thanks for reporting that\r\n\r\nCould you open a PR to fix the features names ?" ]
1,599,857,183,000
1,600,239,559,000
1,600,239,559,000
CONTRIBUTOR
null
null
null
I think the following features in MLQA shouldn't be named the way they are: 1. `questions` (should be `question`) 2. `ids` (should be `id`) 3. `start` (should be `answer_start`) The reasons I'm suggesting these features be renamed are: * To make them consistent with other QA datasets like SQuAD, XQuAD, TyDiQA et...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/619/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/619/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/618
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/618/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/618/comments
https://api.github.com/repos/huggingface/datasets/issues/618/events
https://github.com/huggingface/datasets/pull/618
699,684,831
MDExOlB1bGxSZXF1ZXN0NDg1NDAxMzI5
618
sync logging utils with transformers
{ "login": "stas00", "id": 10676103, "node_id": "MDQ6VXNlcjEwNjc2MTAz", "avatar_url": "https://avatars.githubusercontent.com/u/10676103?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stas00", "html_url": "https://github.com/stas00", "followers_url": "https://api.github.com/users/stas00/fo...
[]
closed
false
null
[]
null
[ "Also, some downloads and dataset processing can be quite long for large datasets like wikipedia/pg19/etc. We probably don't want to user to think that the library is hanging. Happy to reorganize logging between DEBUG/INFO/WARNING to make it less verbose by default though.", "The problem is that `transformers` im...
1,599,853,573,000
1,600,357,259,000
1,600,336,427,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/618", "html_url": "https://github.com/huggingface/datasets/pull/618", "diff_url": "https://github.com/huggingface/datasets/pull/618.diff", "patch_url": "https://github.com/huggingface/datasets/pull/618.patch", "merged_at": null }
sync the docs/code with the recent changes in transformers' `logging` utils: 1. change the default level to `WARNING` 2. add `DATASETS_VERBOSITY` env var 3. expand docs
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/618/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/618/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/617
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/617/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/617/comments
https://api.github.com/repos/huggingface/datasets/issues/617/events
https://github.com/huggingface/datasets/issues/617
699,472,596
MDU6SXNzdWU2OTk0NzI1OTY=
617
Compare different Rouge implementations
{ "login": "ibeltagy", "id": 2287797, "node_id": "MDQ6VXNlcjIyODc3OTc=", "avatar_url": "https://avatars.githubusercontent.com/u/2287797?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ibeltagy", "html_url": "https://github.com/ibeltagy", "followers_url": "https://api.github.com/users/ibelt...
[]
closed
false
null
[]
null
[ "Updates - the differences between the following three\r\n(1) https://github.com/bheinzerling/pyrouge (previously popular. The one I trust the most)\r\n(2) https://github.com/google-research/google-research/tree/master/rouge\r\n(3) https://github.com/pltrdy/files2rouge (used in fairseq)\r\ncan be explained by two t...
1,599,839,372,000
1,617,211,713,000
1,601,632,338,000
NONE
null
null
null
I used RougeL implementation provided in `datasets` [here](https://github.com/huggingface/datasets/blob/master/metrics/rouge/rouge.py) and it gives numbers that match those reported in the pegasus paper but very different from those reported in other papers, [this](https://arxiv.org/pdf/1909.03186.pdf) for example. Ca...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/617/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/617/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/616
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/616/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/616/comments
https://api.github.com/repos/huggingface/datasets/issues/616/events
https://github.com/huggingface/datasets/issues/616
699,462,293
MDU6SXNzdWU2OTk0NjIyOTM=
616
UserWarning: The given NumPy array is not writeable, and PyTorch does not support non-writeable tensors
{ "login": "BramVanroy", "id": 2779410, "node_id": "MDQ6VXNlcjI3Nzk0MTA=", "avatar_url": "https://avatars.githubusercontent.com/u/2779410?v=4", "gravatar_id": "", "url": "https://api.github.com/users/BramVanroy", "html_url": "https://github.com/BramVanroy", "followers_url": "https://api.github.com/users...
[]
open
false
null
[]
null
[ "I have the same issue", "Same issue here when Trying to load a dataset from disk.", "I am also experiencing this issue, and don't know if it's affecting my training.", "Same here. I hope the dataset is not being modified in-place.", "I think the only way to avoid this warning would be to do a copy of the n...
1,599,838,756,000
1,626,988,341,000
null
CONTRIBUTOR
null
null
null
I am trying out the library and want to load in pickled data with `from_dict`. In that dict, one column `text` should be tokenized and the other (an embedding vector) should be retained. All other columns should be removed. When I eventually try to set the format for the columns with `set_format` I am getting this stra...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/616/reactions", "total_count": 4, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 4 }
https://api.github.com/repos/huggingface/datasets/issues/616/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/615
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/615/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/615/comments
https://api.github.com/repos/huggingface/datasets/issues/615/events
https://github.com/huggingface/datasets/issues/615
699,410,773
MDU6SXNzdWU2OTk0MTA3NzM=
615
Offset overflow when slicing a big dataset with an array of indices in Pyarrow >= 1.0.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "Related: https://issues.apache.org/jira/browse/ARROW-9773\r\n\r\nIt's definitely a size thing. I took a smaller dataset with 87000 rows and did:\r\n```\r\nfor i in range(10,1000,20):\r\n table = pa.concat_tables([dset._data]*i)\r\n table.take([0])\r\n```\r\nand it broke at around i=300.\r\n\r\nAlso when `_in...
1,599,835,838,000
1,654,767,648,000
1,600,533,991,000
MEMBER
null
null
null
How to reproduce: ```python from datasets import load_dataset wiki = load_dataset("wikipedia", "20200501.en", split="train") wiki[[0]] --------------------------------------------------------------------------- ArrowInvalid Traceback (most recent call last) <ipython-input-13-38...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/615/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/615/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/614
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/614/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/614/comments
https://api.github.com/repos/huggingface/datasets/issues/614/events
https://github.com/huggingface/datasets/pull/614
699,177,110
MDExOlB1bGxSZXF1ZXN0NDg0OTQ2MzA1
614
[doc] Update deploy.sh
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,599,822,373,000
1,600,073,359,000
1,600,073,357,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/614", "html_url": "https://github.com/huggingface/datasets/pull/614", "diff_url": "https://github.com/huggingface/datasets/pull/614.diff", "patch_url": "https://github.com/huggingface/datasets/pull/614.patch", "merged_at": 1600073357000 }
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/614/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/614/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/613
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/613/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/613/comments
https://api.github.com/repos/huggingface/datasets/issues/613/events
https://github.com/huggingface/datasets/pull/613
699,117,070
MDExOlB1bGxSZXF1ZXN0NDg0ODkyMTUx
613
Add CoNLL-2003 shared task dataset
{ "login": "vblagoje", "id": 458335, "node_id": "MDQ6VXNlcjQ1ODMzNQ==", "avatar_url": "https://avatars.githubusercontent.com/u/458335?v=4", "gravatar_id": "", "url": "https://api.github.com/users/vblagoje", "html_url": "https://github.com/vblagoje", "followers_url": "https://api.github.com/users/vblagoj...
[]
closed
false
null
[]
null
[ "I think we should somewhere mention, that is the dataset in IOB2 tagging scheme, whereas the original dataset uses IOB1 :)", "Indeed this is something we want to mention.\r\n\r\nIf would want to add more details about the IOB1->2 change, feel free to ignore my suggestions and edit the description + update the da...
1,599,818,550,000
1,601,894,585,000
1,600,338,998,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/613", "html_url": "https://github.com/huggingface/datasets/pull/613", "diff_url": "https://github.com/huggingface/datasets/pull/613.diff", "patch_url": "https://github.com/huggingface/datasets/pull/613.patch", "merged_at": 1600338998000 }
Please consider adding CoNLL-2003 shared task dataset as it's beneficial for token classification tasks. The motivation behind this PR is the [PR](https://github.com/huggingface/transformers/pull/7041) in the transformers project. This dataset would be not only useful for the usual run-of-the-mill NER tasks but also fo...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/613/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/613/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/612
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/612/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/612/comments
https://api.github.com/repos/huggingface/datasets/issues/612/events
https://github.com/huggingface/datasets/pull/612
699,008,644
MDExOlB1bGxSZXF1ZXN0NDg0Nzk2Mjg5
612
add multi-proc to dataset dict
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,599,812,293,000
1,599,819,613,000
1,599,819,611,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/612", "html_url": "https://github.com/huggingface/datasets/pull/612", "diff_url": "https://github.com/huggingface/datasets/pull/612.diff", "patch_url": "https://github.com/huggingface/datasets/pull/612.patch", "merged_at": 1599819611000 }
Add multi-proc to `DatasetDict`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/612/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/612/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/611
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/611/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/611/comments
https://api.github.com/repos/huggingface/datasets/issues/611/events
https://github.com/huggingface/datasets/issues/611
698,863,988
MDU6SXNzdWU2OTg4NjM5ODg=
611
ArrowCapacityError: List array cannot contain more than 2147483646 child elements, have 2147483648
{ "login": "sangyx", "id": 32364921, "node_id": "MDQ6VXNlcjMyMzY0OTIx", "avatar_url": "https://avatars.githubusercontent.com/u/32364921?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sangyx", "html_url": "https://github.com/sangyx", "followers_url": "https://api.github.com/users/sangyx/fo...
[]
closed
false
null
[]
null
[ "Can you give us stats/information on your pandas DataFrame?", "```\r\n<class 'pandas.core.frame.DataFrame'>\r\nInt64Index: 17136104 entries, 0 to 17136103\r\nData columns (total 6 columns):\r\n # Column Dtype \r\n--- ------ ----- \r\n 0 item_id int64 \r\n 1 item_titl object \r\n...
1,599,802,152,000
1,654,096,303,000
1,654,096,303,000
NONE
null
null
null
Hi, I'm trying to load a dataset from Dataframe, but I get the error: ```bash --------------------------------------------------------------------------- ArrowCapacityError Traceback (most recent call last) <ipython-input-7-146b6b495963> in <module> ----> 1 dataset = Dataset.from_pandas(emb)...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/611/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/611/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/610
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/610/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/610/comments
https://api.github.com/repos/huggingface/datasets/issues/610/events
https://github.com/huggingface/datasets/issues/610
698,349,388
MDU6SXNzdWU2OTgzNDkzODg=
610
Load text file for RoBERTa pre-training.
{ "login": "chiyuzhang94", "id": 33407613, "node_id": "MDQ6VXNlcjMzNDA3NjEz", "avatar_url": "https://avatars.githubusercontent.com/u/33407613?v=4", "gravatar_id": "", "url": "https://api.github.com/users/chiyuzhang94", "html_url": "https://github.com/chiyuzhang94", "followers_url": "https://api.github.c...
[]
open
false
null
[]
null
[ "Could you try\r\n```python\r\nload_dataset('text', data_files='test.txt',cache_dir=\"./\", split=\"train\")\r\n```\r\n?\r\n\r\n`load_dataset` returns a dictionary by default, like {\"train\": your_dataset}", "Hi @lhoestq\r\nThanks for your suggestion.\r\n\r\nI tried \r\n```\r\ndataset = load_dataset('text', data...
1,599,763,298,000
1,618,044,244,000
null
NONE
null
null
null
I migrate my question from https://github.com/huggingface/transformers/pull/4009#issuecomment-690039444 I tried to train a Roberta from scratch using transformers. But I got OOM issues with loading a large text file. According to the suggestion from @thomwolf , I tried to implement `datasets` to load my text file....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/610/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/610/timeline
null
null
false
https://api.github.com/repos/huggingface/datasets/issues/609
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/609/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/609/comments
https://api.github.com/repos/huggingface/datasets/issues/609/events
https://github.com/huggingface/datasets/pull/609
698,323,989
MDExOlB1bGxSZXF1ZXN0NDg0MTc4Nzky
609
Update GLUE URLs (now hosted on FB)
{ "login": "jeswan", "id": 57466294, "node_id": "MDQ6VXNlcjU3NDY2Mjk0", "avatar_url": "https://avatars.githubusercontent.com/u/57466294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jeswan", "html_url": "https://github.com/jeswan", "followers_url": "https://api.github.com/users/jeswan/fo...
[]
closed
false
null
[]
null
[ "Thanks for opening this PR :) \r\n\r\nWe changed the name of the lib from nlp to datasets yesterday.\r\nCould you rebase from master and re-generate the dataset_info.json file to fix the name changes ?", "Rebased changes here: https://github.com/huggingface/datasets/pull/626" ]
1,599,761,792,000
1,600,110,362,000
1,600,110,361,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/609", "html_url": "https://github.com/huggingface/datasets/pull/609", "diff_url": "https://github.com/huggingface/datasets/pull/609.diff", "patch_url": "https://github.com/huggingface/datasets/pull/609.patch", "merged_at": null }
NYU is switching dataset hosting from Google to FB. This PR closes https://github.com/huggingface/datasets/issues/608 and is necessary for https://github.com/jiant-dev/jiant/issues/161. This PR updates the data URLs based on changes made in https://github.com/nyu-mll/jiant/pull/1112.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/609/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/609/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/608
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/608/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/608/comments
https://api.github.com/repos/huggingface/datasets/issues/608/events
https://github.com/huggingface/datasets/issues/608
698,291,156
MDU6SXNzdWU2OTgyOTExNTY=
608
Don't use the old NYU GLUE dataset URLs
{ "login": "jeswan", "id": 57466294, "node_id": "MDQ6VXNlcjU3NDY2Mjk0", "avatar_url": "https://avatars.githubusercontent.com/u/57466294?v=4", "gravatar_id": "", "url": "https://api.github.com/users/jeswan", "html_url": "https://github.com/jeswan", "followers_url": "https://api.github.com/users/jeswan/fo...
[]
closed
false
null
[]
null
[ "Feel free to open the PR ;)\r\nThanks for updating the dataset_info.json file !" ]
1,599,760,022,000
1,600,239,198,000
1,600,239,198,000
CONTRIBUTOR
null
null
null
NYU is switching dataset hosting from Google to FB. Initial changes to `datasets` are in https://github.com/jeswan/nlp/commit/b7d4a071d432592ded971e30ef73330529de25ce. What tests do you suggest I run before opening a PR? See: https://github.com/jiant-dev/jiant/issues/161 and https://github.com/nyu-mll/jiant/pull/111...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/608/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/608/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/607
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/607/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/607/comments
https://api.github.com/repos/huggingface/datasets/issues/607/events
https://github.com/huggingface/datasets/pull/607
698,094,442
MDExOlB1bGxSZXF1ZXN0NDgzOTcyMDg4
607
Add transmit_format wrapper and tests
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,750,230,000
1,599,751,308,000
1,599,751,307,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/607", "html_url": "https://github.com/huggingface/datasets/pull/607", "diff_url": "https://github.com/huggingface/datasets/pull/607.diff", "patch_url": "https://github.com/huggingface/datasets/pull/607.patch", "merged_at": 1599751307000 }
Same as #605 but using a decorator on-top of dataset transforms that are not in place
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/607/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/607/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/606
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/606/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/606/comments
https://api.github.com/repos/huggingface/datasets/issues/606/events
https://github.com/huggingface/datasets/pull/606
698,050,442
MDExOlB1bGxSZXF1ZXN0NDgzOTMzMDA1
606
Quick fix :)
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ ":heart:" ]
1,599,748,326,000
1,599,754,712,000
1,599,754,710,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/606", "html_url": "https://github.com/huggingface/datasets/pull/606", "diff_url": "https://github.com/huggingface/datasets/pull/606.diff", "patch_url": "https://github.com/huggingface/datasets/pull/606.patch", "merged_at": 1599754710000 }
`nlp` => `datasets`
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/606/reactions", "total_count": 3, "+1": 0, "-1": 0, "laugh": 1, "hooray": 1, "confused": 0, "heart": 1, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/606/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/605
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/605/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/605/comments
https://api.github.com/repos/huggingface/datasets/issues/605/events
https://github.com/huggingface/datasets/pull/605
697,887,401
MDExOlB1bGxSZXF1ZXN0NDgzNzg1Mjc1
605
[Datasets] Transmit format to children
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Closing as #607 was merged" ]
1,599,741,018,000
1,599,754,521,000
1,599,754,521,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/605", "html_url": "https://github.com/huggingface/datasets/pull/605", "diff_url": "https://github.com/huggingface/datasets/pull/605.diff", "patch_url": "https://github.com/huggingface/datasets/pull/605.patch", "merged_at": null }
Transmit format to children obtained when processing a dataset. Added a test. When concatenating datasets, if the formats are disparate, the concatenated dataset has a format reset to defaults.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/605/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/605/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/604
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/604/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/604/comments
https://api.github.com/repos/huggingface/datasets/issues/604/events
https://github.com/huggingface/datasets/pull/604
697,774,581
MDExOlB1bGxSZXF1ZXN0NDgzNjgxNTc0
604
Update bucket prefix
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,735,673,000
1,599,741,933,000
1,599,741,932,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/604", "html_url": "https://github.com/huggingface/datasets/pull/604", "diff_url": "https://github.com/huggingface/datasets/pull/604.diff", "patch_url": "https://github.com/huggingface/datasets/pull/604.patch", "merged_at": 1599741932000 }
cc @julien-c
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/604/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/604/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/603
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/603/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/603/comments
https://api.github.com/repos/huggingface/datasets/issues/603/events
https://github.com/huggingface/datasets/pull/603
697,758,750
MDExOlB1bGxSZXF1ZXN0NDgzNjY2ODk5
603
Set scripts version to master
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,734,864,000
1,599,735,725,000
1,599,735,724,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/603", "html_url": "https://github.com/huggingface/datasets/pull/603", "diff_url": "https://github.com/huggingface/datasets/pull/603.diff", "patch_url": "https://github.com/huggingface/datasets/pull/603.patch", "merged_at": 1599735724000 }
By default the scripts version is master, so that if the library is installed with ``` pip install git+http://github.com/huggingface/nlp.git ``` or ``` git clone http://github.com/huggingface/nlp.git pip install -e ./nlp ``` will use the latest scripts, and not the ones from the previous version.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/603/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/603/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/602
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/602/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/602/comments
https://api.github.com/repos/huggingface/datasets/issues/602/events
https://github.com/huggingface/datasets/pull/602
697,636,605
MDExOlB1bGxSZXF1ZXN0NDgzNTU3NDM0
602
apply offset to indices in multiprocessed map
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,728,070,000
1,599,735,819,000
1,599,735,817,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/602", "html_url": "https://github.com/huggingface/datasets/pull/602", "diff_url": "https://github.com/huggingface/datasets/pull/602.diff", "patch_url": "https://github.com/huggingface/datasets/pull/602.patch", "merged_at": 1599735817000 }
Fix #597 I fixed the indices by applying an offset. I added the case to our tests to make sure it doesn't happen again. I also added the message proposed by @thomwolf in #597 ```python >>> d.select(range(10)).map(fn, with_indices=True, batched=True, num_proc=2, load_from_cache_file=False) Done writing 10 ...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/602/reactions", "total_count": 1, "+1": 1, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/602/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/601
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/601/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/601/comments
https://api.github.com/repos/huggingface/datasets/issues/601/events
https://github.com/huggingface/datasets/pull/601
697,574,848
MDExOlB1bGxSZXF1ZXN0NDgzNTAzMjAw
601
check if trasnformers has PreTrainedTokenizerBase
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,724,496,000
1,599,735,697,000
1,599,735,696,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/601", "html_url": "https://github.com/huggingface/datasets/pull/601", "diff_url": "https://github.com/huggingface/datasets/pull/601.diff", "patch_url": "https://github.com/huggingface/datasets/pull/601.patch", "merged_at": 1599735696000 }
Fix #598
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/601/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/601/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/600
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/600/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/600/comments
https://api.github.com/repos/huggingface/datasets/issues/600/events
https://github.com/huggingface/datasets/issues/600
697,496,913
MDU6SXNzdWU2OTc0OTY5MTM=
600
Pickling error when loading dataset
{ "login": "kandorm", "id": 17310286, "node_id": "MDQ6VXNlcjE3MzEwMjg2", "avatar_url": "https://avatars.githubusercontent.com/u/17310286?v=4", "gravatar_id": "", "url": "https://api.github.com/users/kandorm", "html_url": "https://github.com/kandorm", "followers_url": "https://api.github.com/users/kandor...
[]
closed
false
null
[]
null
[ "When I change from python3.6 to python3.8, it works! ", "Does it work when you install `nlp` from source on python 3.6?", "No, still the pickling error.", "I wasn't able to reproduce on google colab (python 3.6.9 as well) with \r\n\r\npickle==4.0\r\ndill=0.3.2\r\ntransformers==3.1.0\r\ndatasets=1.0.1 (also t...
1,599,719,288,000
1,601,044,314,000
1,601,044,314,000
NONE
null
null
null
Hi, I modified line 136 in the original [run_language_modeling.py](https://github.com/huggingface/transformers/blob/master/examples/language-modeling/run_language_modeling.py) as: ``` # line 136: return LineByLineTextDataset(tokenizer=tokenizer, file_path=file_path, block_size=args.block_size) dataset = load_da...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/600/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/600/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/599
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/599/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/599/comments
https://api.github.com/repos/huggingface/datasets/issues/599/events
https://github.com/huggingface/datasets/pull/599
697,377,786
MDExOlB1bGxSZXF1ZXN0NDgzMzI3ODQ5
599
Add MATINF dataset
{ "login": "JetRunner", "id": 22514219, "node_id": "MDQ6VXNlcjIyNTE0MjE5", "avatar_url": "https://avatars.githubusercontent.com/u/22514219?v=4", "gravatar_id": "", "url": "https://api.github.com/users/JetRunner", "html_url": "https://github.com/JetRunner", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi ! sorry for the late response\r\n\r\nCould you try to rebase from master ? We changed the named of the library last week so you have to include this change in your code.\r\n\r\nCan you give me more details about the error you get when running the cli command ?\r\n\r\nNote that in case of a manual download you h...
1,599,708,669,000
1,600,345,045,000
1,600,345,045,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/599", "html_url": "https://github.com/huggingface/datasets/pull/599", "diff_url": "https://github.com/huggingface/datasets/pull/599.diff", "patch_url": "https://github.com/huggingface/datasets/pull/599.patch", "merged_at": null }
@lhoestq The command to create metadata failed. I guess it's because the zip is not downloaded from a remote address? How to solve that? Also the CI fails and I don't know how to fix that :(
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/599/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/599/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/598
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/598/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/598/comments
https://api.github.com/repos/huggingface/datasets/issues/598/events
https://github.com/huggingface/datasets/issues/598
697,156,501
MDU6SXNzdWU2OTcxNTY1MDE=
598
The current version of the package on github has an error when loading dataset
{ "login": "zeyuyun1", "id": 43428393, "node_id": "MDQ6VXNlcjQzNDI4Mzkz", "avatar_url": "https://avatars.githubusercontent.com/u/43428393?v=4", "gravatar_id": "", "url": "https://api.github.com/users/zeyuyun1", "html_url": "https://github.com/zeyuyun1", "followers_url": "https://api.github.com/users/zey...
[]
closed
false
null
[]
null
[ "Thanks for reporting !\r\nWhich version of transformers are you using ?\r\nIt looks like it doesn't have the PreTrainedTokenizerBase class", "I was using transformer 2.9. And I switch to the latest transformer package. Everything works just fine!!\r\n\r\nThanks for helping! I should look more carefully next time...
1,599,685,403,000
1,599,719,121,000
1,599,692,248,000
NONE
null
null
null
Instead of downloading the package from pip, downloading the version from source will result in an error when loading dataset (the pip version is completely fine): To recreate the error: First, installing nlp directly from source: ``` git clone https://github.com/huggingface/nlp.git cd nlp pip install -e . ``...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/598/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/598/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/597
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/597/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/597/comments
https://api.github.com/repos/huggingface/datasets/issues/597/events
https://github.com/huggingface/datasets/issues/597
697,112,029
MDU6SXNzdWU2OTcxMTIwMjk=
597
Indices incorrect with multiprocessing
{ "login": "joeddav", "id": 9353833, "node_id": "MDQ6VXNlcjkzNTM4MzM=", "avatar_url": "https://avatars.githubusercontent.com/u/9353833?v=4", "gravatar_id": "", "url": "https://api.github.com/users/joeddav", "html_url": "https://github.com/joeddav", "followers_url": "https://api.github.com/users/joeddav/...
[]
closed
false
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[ { "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.git...
null
[ "I fixed a bug that could cause this issue earlier today. Could you pull the latest version and try again ?", "Still the case on master.\r\nI guess we should have an offset in the multi-procs indeed (hopefully it's enough).\r\n\r\nAlso, side note is that we should add some logging before the \"test\" to say we ar...
1,599,681,056,000
1,599,735,817,000
1,599,735,817,000
CONTRIBUTOR
null
null
null
When `num_proc` > 1, the indices argument passed to the map function is incorrect: ```python d = load_dataset('imdb', split='test[:1%]') def fn(x, inds): print(inds) return x d.select(range(10)).map(fn, with_indices=True, batched=True) # [0, 1] # [0, 1, 2, 3, 4, 5, 6, 7, 8, 9] d.select(range(10...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/597/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/597/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/596
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/596/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/596/comments
https://api.github.com/repos/huggingface/datasets/issues/596/events
https://github.com/huggingface/datasets/pull/596
696,928,139
MDExOlB1bGxSZXF1ZXN0NDgyOTM5MTgw
596
[style/quality] Moving to isort 5.0.0 + style/quality on datasets and metrics
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[ "Ready for review @lhoestq, just updated a few 156 files here" ]
1,599,666,441,000
1,599,732,304,000
1,599,732,303,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/596", "html_url": "https://github.com/huggingface/datasets/pull/596", "diff_url": "https://github.com/huggingface/datasets/pull/596.diff", "patch_url": "https://github.com/huggingface/datasets/pull/596.patch", "merged_at": 1599732303000 }
Move the repo to isort 5.0.0. Also start testing style/quality on datasets and metrics. Specific rule: we allow F401 (unused imports) in metrics to be able to add imports to detect early on missing dependencies. Maybe we could add this in datasets but while cleaning this I've seen many example of really unused i...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/596/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/596/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/595
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/595/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/595/comments
https://api.github.com/repos/huggingface/datasets/issues/595/events
https://github.com/huggingface/datasets/issues/595
696,892,304
MDU6SXNzdWU2OTY4OTIzMDQ=
595
`Dataset`/`DatasetDict` has no attribute 'save_to_disk'
{ "login": "sudarshan85", "id": 488428, "node_id": "MDQ6VXNlcjQ4ODQyOA==", "avatar_url": "https://avatars.githubusercontent.com/u/488428?v=4", "gravatar_id": "", "url": "https://api.github.com/users/sudarshan85", "html_url": "https://github.com/sudarshan85", "followers_url": "https://api.github.com/user...
[]
closed
false
null
[]
null
[ "`pip install git+https://github.com/huggingface/nlp.git` should have done the job.\r\n\r\nDid you uninstall `nlp` before installing from github ?", "> Did you uninstall `nlp` before installing from github ?\r\n\r\nI did not. I created a new environment and installed `nlp` directly from `github` and it worked!\r\...
1,599,663,712,000
1,599,668,419,000
1,599,668,418,000
NONE
null
null
null
Hi, As the title indicates, both `Dataset` and `DatasetDict` classes don't seem to have the `save_to_disk` method. While the file [`arrow_dataset.py`](https://github.com/huggingface/nlp/blob/34bf0b03bfe03e7f77b8fec1cd48f5452c4fc7c1/src/nlp/arrow_dataset.py) in the repo here has the method, the file `arrow_dataset.p...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/595/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/595/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/594
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/594/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/594/comments
https://api.github.com/repos/huggingface/datasets/issues/594/events
https://github.com/huggingface/datasets/pull/594
696,816,893
MDExOlB1bGxSZXF1ZXN0NDgyODQ1OTc5
594
Fix germeval url
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,658,175,000
1,599,658,475,000
1,599,658,474,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/594", "html_url": "https://github.com/huggingface/datasets/pull/594", "diff_url": "https://github.com/huggingface/datasets/pull/594.diff", "patch_url": "https://github.com/huggingface/datasets/pull/594.patch", "merged_at": 1599658474000 }
Continuation of #593 but without the dummy data hack
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/594/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/594/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/593
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/593/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/593/comments
https://api.github.com/repos/huggingface/datasets/issues/593/events
https://github.com/huggingface/datasets/pull/593
696,679,182
MDExOlB1bGxSZXF1ZXN0NDgyNzI5NTgw
593
GermEval 2014: new download urls
{ "login": "stefan-it", "id": 20651387, "node_id": "MDQ6VXNlcjIwNjUxMzg3", "avatar_url": "https://avatars.githubusercontent.com/u/20651387?v=4", "gravatar_id": "", "url": "https://api.github.com/users/stefan-it", "html_url": "https://github.com/stefan-it", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "/cc: @vblagoje", "Closing this one as #594 is merged (same changes except the dummy data hack)", "Awesome @stefan-it ! @lhoestq how soon can I use the fixed GermEval dataset in HF token classification examples?", "I've manually updated the script on S3, so you can actually use it right now with\r\n```python\...
1,599,646,049,000
1,599,661,014,000
1,599,658,515,000
CONTRIBUTOR
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/593", "html_url": "https://github.com/huggingface/datasets/pull/593", "diff_url": "https://github.com/huggingface/datasets/pull/593.diff", "patch_url": "https://github.com/huggingface/datasets/pull/593.patch", "merged_at": null }
Hi, unfortunately, the download links for the GermEval 2014 dataset have changed: they're now located on a Google Drive. I changed the URLs and bump version from 1.0.0 to 2.0.0.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/593/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/593/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/592
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/592/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/592/comments
https://api.github.com/repos/huggingface/datasets/issues/592/events
https://github.com/huggingface/datasets/pull/592
696,619,986
MDExOlB1bGxSZXF1ZXN0NDgyNjc4MDkw
592
Test in memory and on disk
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,641,970,000
1,599,659,404,000
1,599,659,403,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/592", "html_url": "https://github.com/huggingface/datasets/pull/592", "diff_url": "https://github.com/huggingface/datasets/pull/592.diff", "patch_url": "https://github.com/huggingface/datasets/pull/592.patch", "merged_at": 1599659403000 }
I added test parameters to do every test both in memory and on disk. I also found a bug in concatenate_dataset thanks to the new tests and fixed it.
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/592/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/592/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/591
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/591/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/591/comments
https://api.github.com/repos/huggingface/datasets/issues/591/events
https://github.com/huggingface/datasets/pull/591
696,530,413
MDExOlB1bGxSZXF1ZXN0NDgyNjAxMzc1
591
fix #589 (backward compat)
{ "login": "thomwolf", "id": 7353373, "node_id": "MDQ6VXNlcjczNTMzNzM=", "avatar_url": "https://avatars.githubusercontent.com/u/7353373?v=4", "gravatar_id": "", "url": "https://api.github.com/users/thomwolf", "html_url": "https://github.com/thomwolf", "followers_url": "https://api.github.com/users/thomw...
[]
closed
false
null
[]
null
[]
1,599,636,793,000
1,599,641,876,000
1,599,641,875,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/591", "html_url": "https://github.com/huggingface/datasets/pull/591", "diff_url": "https://github.com/huggingface/datasets/pull/591.diff", "patch_url": "https://github.com/huggingface/datasets/pull/591.patch", "merged_at": 1599641874000 }
Fix #589
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/591/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/591/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/590
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/590/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/590/comments
https://api.github.com/repos/huggingface/datasets/issues/590/events
https://github.com/huggingface/datasets/issues/590
696,501,827
MDU6SXNzdWU2OTY1MDE4Mjc=
590
The process cannot access the file because it is being used by another process (windows)
{ "login": "saareliad", "id": 22762845, "node_id": "MDQ6VXNlcjIyNzYyODQ1", "avatar_url": "https://avatars.githubusercontent.com/u/22762845?v=4", "gravatar_id": "", "url": "https://api.github.com/users/saareliad", "html_url": "https://github.com/saareliad", "followers_url": "https://api.github.com/users/...
[]
closed
false
null
[]
null
[ "Hi, which version of `nlp` are you using?\r\n\r\nBy the way we'll be releasing today a significant update fixing many issues (but also comprising a few breaking changes).\r\nYou can see more informations here #545 and try it by installing from source from the master branch.", "I'm using version 0.4.0.\r\n\r\n", ...
1,599,634,896,000
1,601,042,548,000
1,601,042,548,000
NONE
null
null
null
Hi, I consistently get the following error when developing in my PC (windows 10): ``` train_dataset = train_dataset.map(convert_to_features, batched=True) File "C:\Users\saareliad\AppData\Local\Continuum\miniconda3\envs\py38\lib\site-packages\nlp\arrow_dataset.py", line 970, in map shutil.move(tmp_file....
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/590/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 1 }
https://api.github.com/repos/huggingface/datasets/issues/590/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/589
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/589/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/589/comments
https://api.github.com/repos/huggingface/datasets/issues/589/events
https://github.com/huggingface/datasets/issues/589
696,488,447
MDU6SXNzdWU2OTY0ODg0NDc=
589
Cannot use nlp.load_dataset text, AttributeError: module 'nlp.utils' has no attribute 'logging'
{ "login": "ksjae", "id": 17930170, "node_id": "MDQ6VXNlcjE3OTMwMTcw", "avatar_url": "https://avatars.githubusercontent.com/u/17930170?v=4", "gravatar_id": "", "url": "https://api.github.com/users/ksjae", "html_url": "https://github.com/ksjae", "followers_url": "https://api.github.com/users/ksjae/follow...
[]
closed
false
null
[]
null
[]
1,599,634,013,000
1,599,641,874,000
1,599,641,874,000
NONE
null
null
null
``` Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/root/anaconda3/envs/pytorch/lib/python3.7/site-packages/nlp/load.py", line 533, in load_dataset builder_cls = import_main_class(module_path, dataset=True) File "/root/anaconda3/envs/pytorch/lib/python3.7/site-packages/nlp...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/589/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/589/timeline
null
completed
false
https://api.github.com/repos/huggingface/datasets/issues/588
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/588/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/588/comments
https://api.github.com/repos/huggingface/datasets/issues/588/events
https://github.com/huggingface/datasets/pull/588
695,249,809
MDExOlB1bGxSZXF1ZXN0NDgxNTE5NzQx
588
Support pathlike obj in load dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,495,201,000
1,599,551,119,000
1,599,551,118,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/588", "html_url": "https://github.com/huggingface/datasets/pull/588", "diff_url": "https://github.com/huggingface/datasets/pull/588.diff", "patch_url": "https://github.com/huggingface/datasets/pull/588.patch", "merged_at": 1599551117000 }
Fix #582 (I recreated the PR, I got an issue with git)
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/588/reactions", "total_count": 1, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 1, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/588/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/587
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/587/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/587/comments
https://api.github.com/repos/huggingface/datasets/issues/587/events
https://github.com/huggingface/datasets/pull/587
695,246,018
MDExOlB1bGxSZXF1ZXN0NDgxNTE2Mzkx
587
Support pathlike obj in load dataset
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,494,956,000
1,599,495,035,000
1,599,495,035,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/587", "html_url": "https://github.com/huggingface/datasets/pull/587", "diff_url": "https://github.com/huggingface/datasets/pull/587.diff", "patch_url": "https://github.com/huggingface/datasets/pull/587.patch", "merged_at": null }
Fix #582
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/587/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/587/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/586
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/586/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/586/comments
https://api.github.com/repos/huggingface/datasets/issues/586/events
https://github.com/huggingface/datasets/pull/586
695,237,999
MDExOlB1bGxSZXF1ZXN0NDgxNTA5MzU1
586
Better message when data files is empty
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,494,397,000
1,599,642,009,000
1,599,642,008,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/586", "html_url": "https://github.com/huggingface/datasets/pull/586", "diff_url": "https://github.com/huggingface/datasets/pull/586.diff", "patch_url": "https://github.com/huggingface/datasets/pull/586.patch", "merged_at": 1599642007000 }
Fix #581
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/586/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/586/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/585
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/585/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/585/comments
https://api.github.com/repos/huggingface/datasets/issues/585/events
https://github.com/huggingface/datasets/pull/585
695,191,209
MDExOlB1bGxSZXF1ZXN0NDgxNDY4NTM4
585
Fix select for pyarrow < 1.0.0
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[]
1,599,490,972,000
1,599,550,997,000
1,599,550,995,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/585", "html_url": "https://github.com/huggingface/datasets/pull/585", "diff_url": "https://github.com/huggingface/datasets/pull/585.diff", "patch_url": "https://github.com/huggingface/datasets/pull/585.patch", "merged_at": 1599550995000 }
Fix #583
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/585/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/585/timeline
null
null
true
https://api.github.com/repos/huggingface/datasets/issues/584
https://api.github.com/repos/huggingface/datasets
https://api.github.com/repos/huggingface/datasets/issues/584/labels{/name}
https://api.github.com/repos/huggingface/datasets/issues/584/comments
https://api.github.com/repos/huggingface/datasets/issues/584/events
https://github.com/huggingface/datasets/pull/584
695,186,652
MDExOlB1bGxSZXF1ZXN0NDgxNDY0NjEz
584
Use github versioning
{ "login": "lhoestq", "id": 42851186, "node_id": "MDQ6VXNlcjQyODUxMTg2", "avatar_url": "https://avatars.githubusercontent.com/u/42851186?v=4", "gravatar_id": "", "url": "https://api.github.com/users/lhoestq", "html_url": "https://github.com/lhoestq", "followers_url": "https://api.github.com/users/lhoest...
[]
closed
false
null
[]
null
[ "I noticed that datasets like `cnn_dailymail` need the `version` parameter to be passed to its `config_kwargs`.\r\nShall we rename the `version` paramater in `load_dataset` ? Maybe `repo_version` or `script_version` ?" ]
1,599,490,695,000
1,599,658,655,000
1,599,658,654,000
MEMBER
null
false
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/584", "html_url": "https://github.com/huggingface/datasets/pull/584", "diff_url": "https://github.com/huggingface/datasets/pull/584.diff", "patch_url": "https://github.com/huggingface/datasets/pull/584.patch", "merged_at": 1599658654000 }
Right now dataset scripts and metrics are downloaded from S3 which is in sync with master. It means that it's not currently possible to pin the dataset/metric script version. To fix that I changed the download url from S3 to github, and adding a `version` parameter in `load_dataset` and `load_metric` to pin a certai...
{ "url": "https://api.github.com/repos/huggingface/datasets/issues/584/reactions", "total_count": 0, "+1": 0, "-1": 0, "laugh": 0, "hooray": 0, "confused": 0, "heart": 0, "rocket": 0, "eyes": 0 }
https://api.github.com/repos/huggingface/datasets/issues/584/timeline
null
null
true