Dataset Viewer
Auto-converted to Parquet Duplicate
html_url
stringlengths
48
51
title
stringlengths
5
268
comments
stringlengths
70
51.8k
body
stringlengths
0
29.8k
comment_length
int64
16
1.52k
text
stringlengths
164
54.1k
embeddings
sequence
https://github.com/huggingface/datasets/issues/2945
Protect master branch
@lhoestq now the 2 are implemented. Please note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to...
After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.: - 00cc036fea7c7745cfe722360036ed306796a3f2 - 13ae8c98602bbad8197de3b9b425f4c78f582af1 - ... I propo...
64
Protect master branch After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.: - 00cc036fea7c7745cfe722360036ed306796a3f2 - 13ae8c98602bbad8197de3b9b425f4c78f...
[ -0.15531952679157257, -0.1002306342124939, -0.0703212171792984, -0.07981692999601364, -0.10425391793251038, -0.18798768520355225, 0.010403447784483433, 0.2728639543056488, -0.009848886169493198, -0.08413966745138168, 0.2926151156425476, -0.07787521183490753, -0.1324571818113327, 0.21589028...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`. To avoid other users from having this issue we could make the caching differentiate the two, what do you think ?
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
50
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499544739723206, 0.12438808381557465, -0.04655871540307999, 0.23689685761928558, 0.15172921121120453, -0.06666013598442078, -0.006941257044672966, 0.3285394608974457, 0.17373836040496826, 0.029381630942225456, -0.2313861846923828, 0.3581928610801697, -0.1503300666809082, 0.25273647904...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
If it's easy enough to implement, then yes please 😄 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests.
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
28
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499544739723206, 0.12438808381557465, -0.04655871540307999, 0.23689685761928558, 0.15172921121120453, -0.06666013598442078, -0.006941257044672966, 0.3285394608974457, 0.17373836040496826, 0.029381630942225456, -0.2313861846923828, 0.3581928610801697, -0.1503300666809082, 0.25273647904...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
22
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499544739723206, 0.12438808381557465, -0.04655871540307999, 0.23689685761928558, 0.15172921121120453, -0.06666013598442078, -0.006941257044672966, 0.3285394608974457, 0.17373836040496826, 0.029381630942225456, -0.2313861846923828, 0.3581928610801697, -0.1503300666809082, 0.25273647904...
https://github.com/huggingface/datasets/issues/2943
Backwards compatibility broken for cached datasets that use `.filter()`
I just merged a fix, let me know if you're still having this kind of issues :) We'll do a release soon to make this fix available
## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in...
27
Backwards compatibility broken for cached datasets that use `.filter()` ## Describe the bug After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with `ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No...
[ -0.30499544739723206, 0.12438808381557465, -0.04655871540307999, 0.23689685761928558, 0.15172921121120453, -0.06666013598442078, -0.006941257044672966, 0.3285394608974457, 0.17373836040496826, 0.029381630942225456, -0.2313861846923828, 0.3581928610801697, -0.1503300666809082, 0.25273647904...
https://github.com/huggingface/datasets/issues/2937
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
Hi @daqieq, thanks for reporting. Unfortunately, I was not able to reproduce this bug: ```ipython In [1]: from datasets import load_dataset ...: ds = load_dataset('wiki_bio') Downloading: 7.58kB [00:00, 26.3kB/s] Downloading: 2.71kB [00:00, ?B/s] Using custom data configuration default Downloading and prep...
## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset('wiki_bio') ``` ## Expected results It is expected that the dataset downloads without any er...
109
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied ## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_datas...
[ -0.22893071174621582, 0.38345417380332947, 0.04034576937556267, 0.25503984093666077, -0.012111208401620388, 0.2622571289539337, 0.5036684274673462, 0.13295303285121918, 0.34428197145462036, 0.1357172727584839, -0.10897453129291534, 0.07728376984596252, 0.07315050810575485, -0.0056993099860...
https://github.com/huggingface/datasets/issues/2937
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine. Running on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename an...
## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_dataset('wiki_bio') ``` ## Expected results It is expected that the dataset downloads without any er...
194
load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied ## Describe the bug Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11. ## Steps to reproduce the bug ```python from datasets import load_dataset ds = load_datas...
[ -0.22893071174621582, 0.38345417380332947, 0.04034576937556267, 0.25503984093666077, -0.012111208401620388, 0.2622571289539337, 0.5036684274673462, 0.13295303285121918, 0.34428197145462036, 0.1357172727584839, -0.10897453129291534, 0.07728376984596252, 0.07315050810575485, -0.0056993099860...
https://github.com/huggingface/datasets/issues/2934
to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows
I did some investigation and, as it seems, the bug stems from [this line](https://github.com/huggingface/datasets/blob/8004d7c3e1d74b29c3e5b0d1660331cd26758363/src/datasets/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution...
To reproduce: ```python import datasets as ds import weakref import gc d = ds.load_dataset("mnist", split="train") ref = weakref.ref(d._data.table) tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="label") del tfd, d gc.collect() assert ref() is None, "Error: there is at least one refe...
99
to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows To reproduce: ```python import datasets as ds import weakref import gc d = ds.load_dataset("mnist", split="train") ref = weakref.ref(d._data.table) tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="lab...
[ 0.04569048434495926, 0.3345080316066742, 0.1148887500166893, 0.08045100420713425, 0.23043039441108704, 0.11888763308525085, 0.3988848924636841, 0.2548752427101135, -0.10991697013378143, 0.26743850111961365, -0.3353854715824127, 0.4040723741054535, -0.16448107361793518, -0.1074628010392189,...
https://github.com/huggingface/datasets/issues/2924
"File name too long" error for file locks
"Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_d(...TRUNCATED)
"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED)
39
"\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED)
[0.04933885484933853,0.0920017659664154,-0.07149408012628555,0.40184855461120605,0.4151674211025238,(...TRUNCATED)
https://github.com/huggingface/datasets/issues/2924
"File name too long" error for file locks
"Yes, you're right! I need to get you more info here. Either there's something going with the name i(...TRUNCATED)
"## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED)
67
"\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED)
[0.04933885484933853,0.0920017659664154,-0.07149408012628555,0.40184855461120605,0.4151674211025238,(...TRUNCATED)
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
12