html_url stringlengths 48 51 | title stringlengths 5 268 | comments stringlengths 70 51.8k | body stringlengths 0 29.8k | comment_length int64 16 1.52k | text stringlengths 164 54.1k | embeddings sequence |
|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/2945 | Protect master branch | @lhoestq now the 2 are implemented.
Please note that for the the second protection, finally I have chosen to protect the master branch only from **merge commits** (see update comment above), so no need to disable/re-enable the protection on each release (direct commits, different from merge commits, can be pushed to... | After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:
- 00cc036fea7c7745cfe722360036ed306796a3f2
- 13ae8c98602bbad8197de3b9b425f4c78f582af1
- ...
I propo... | 64 | Protect master branch
After accidental merge commit (91c55355b634d0dc73350a7ddee1a6776dbbdd69) into `datasets` master branch, all commits present in the feature branch were permanently added to `datasets` master branch history, as e.g.:
- 00cc036fea7c7745cfe722360036ed306796a3f2
- 13ae8c98602bbad8197de3b9b425f4c78f... | [
-0.15531952679157257,
-0.1002306342124939,
-0.0703212171792984,
-0.07981692999601364,
-0.10425391793251038,
-0.18798768520355225,
0.010403447784483433,
0.2728639543056488,
-0.009848886169493198,
-0.08413966745138168,
0.2926151156425476,
-0.07787521183490753,
-0.1324571818113327,
0.21589028... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | Hi ! I guess the caching mechanism should have considered the new `filter` to be different from the old one, and don't use cached results from the old `filter`.
To avoid other users from having this issue we could make the caching differentiate the two, what do you think ? | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 50 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499544739723206,
0.12438808381557465,
-0.04655871540307999,
0.23689685761928558,
0.15172921121120453,
-0.06666013598442078,
-0.006941257044672966,
0.3285394608974457,
0.17373836040496826,
0.029381630942225456,
-0.2313861846923828,
0.3581928610801697,
-0.1503300666809082,
0.25273647904... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | If it's easy enough to implement, then yes please 😄 But this issue can be low-priority, since I've only encountered it in a couple of `transformers` CI tests. | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 28 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499544739723206,
0.12438808381557465,
-0.04655871540307999,
0.23689685761928558,
0.15172921121120453,
-0.06666013598442078,
-0.006941257044672966,
0.3285394608974457,
0.17373836040496826,
0.029381630942225456,
-0.2313861846923828,
0.3581928610801697,
-0.1503300666809082,
0.25273647904... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | Well it can cause issue with anyone that updates `datasets` and re-run some code that uses filter, so I'm creating a PR | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 22 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499544739723206,
0.12438808381557465,
-0.04655871540307999,
0.23689685761928558,
0.15172921121120453,
-0.06666013598442078,
-0.006941257044672966,
0.3285394608974457,
0.17373836040496826,
0.029381630942225456,
-0.2313861846923828,
0.3581928610801697,
-0.1503300666809082,
0.25273647904... |
https://github.com/huggingface/datasets/issues/2943 | Backwards compatibility broken for cached datasets that use `.filter()` | I just merged a fix, let me know if you're still having this kind of issues :)
We'll do a release soon to make this fix available | ## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=None), 'text': Value(dtype='string', id=None), 'speaker_id': Value(dtype='in... | 27 | Backwards compatibility broken for cached datasets that use `.filter()`
## Describe the bug
After upgrading to datasets `1.12.0`, some cached `.filter()` steps from `1.11.0` started failing with
`ValueError: Keys mismatch: between {'indices': Value(dtype='uint64', id=None)} and {'file': Value(dtype='string', id=No... | [
-0.30499544739723206,
0.12438808381557465,
-0.04655871540307999,
0.23689685761928558,
0.15172921121120453,
-0.06666013598442078,
-0.006941257044672966,
0.3285394608974457,
0.17373836040496826,
0.029381630942225456,
-0.2313861846923828,
0.3581928610801697,
-0.1503300666809082,
0.25273647904... |
https://github.com/huggingface/datasets/issues/2937 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied | Hi @daqieq, thanks for reporting.
Unfortunately, I was not able to reproduce this bug:
```ipython
In [1]: from datasets import load_dataset
...: ds = load_dataset('wiki_bio')
Downloading: 7.58kB [00:00, 26.3kB/s]
Downloading: 2.71kB [00:00, ?B/s]
Using custom data configuration default
Downloading and prep... | ## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset('wiki_bio')
```
## Expected results
It is expected that the dataset downloads without any er... | 109 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_datas... | [
-0.22893071174621582,
0.38345417380332947,
0.04034576937556267,
0.25503984093666077,
-0.012111208401620388,
0.2622571289539337,
0.5036684274673462,
0.13295303285121918,
0.34428197145462036,
0.1357172727584839,
-0.10897453129291534,
0.07728376984596252,
0.07315050810575485,
-0.0056993099860... |
https://github.com/huggingface/datasets/issues/2937 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied | Thanks @albertvillanova for looking at it! I tried on my personal Windows machine and it downloaded just fine.
Running on my work machine and on a colleague's machine it is consistently hitting this error. It's not a write access issue because the `.incomplete` directory is written just fine. It just won't rename an... | ## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset('wiki_bio')
```
## Expected results
It is expected that the dataset downloads without any er... | 194 | load_dataset using default cache on Windows causes PermissionError: [WinError 5] Access is denied
## Describe the bug
Standard process to download and load the wiki_bio dataset causes PermissionError in Windows 10 and 11.
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_datas... | [
-0.22893071174621582,
0.38345417380332947,
0.04034576937556267,
0.25503984093666077,
-0.012111208401620388,
0.2622571289539337,
0.5036684274673462,
0.13295303285121918,
0.34428197145462036,
0.1357172727584839,
-0.10897453129291534,
0.07728376984596252,
0.07315050810575485,
-0.0056993099860... |
https://github.com/huggingface/datasets/issues/2934 | to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows | I did some investigation and, as it seems, the bug stems from [this line](https://github.com/huggingface/datasets/blob/8004d7c3e1d74b29c3e5b0d1660331cd26758363/src/datasets/arrow_dataset.py#L325). The lifecycle of the dataset from the linked line is bound to one of the returned `tf.data.Dataset`. So my (hacky) solution... | To reproduce:
```python
import datasets as ds
import weakref
import gc
d = ds.load_dataset("mnist", split="train")
ref = weakref.ref(d._data.table)
tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="label")
del tfd, d
gc.collect()
assert ref() is None, "Error: there is at least one refe... | 99 | to_tf_dataset keeps a reference to the open data somewhere, causing issues on windows
To reproduce:
```python
import datasets as ds
import weakref
import gc
d = ds.load_dataset("mnist", split="train")
ref = weakref.ref(d._data.table)
tfd = d.to_tf_dataset("image", batch_size=1, shuffle=False, label_cols="lab... | [
0.04569048434495926,
0.3345080316066742,
0.1148887500166893,
0.08045100420713425,
0.23043039441108704,
0.11888763308525085,
0.3988848924636841,
0.2548752427101135,
-0.10991697013378143,
0.26743850111961365,
-0.3353854715824127,
0.4040723741054535,
-0.16448107361793518,
-0.1074628010392189,... |
https://github.com/huggingface/datasets/issues/2924 | "File name too long" error for file locks | "Hi, the filename here is less than 255\r\n```python\r\n>>> len(\"_home_garrett_.cache_huggingface_d(...TRUNCATED) | "## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED) | 39 | "\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED) | [0.04933885484933853,0.0920017659664154,-0.07149408012628555,0.40184855461120605,0.4151674211025238,(...TRUNCATED) |
https://github.com/huggingface/datasets/issues/2924 | "File name too long" error for file locks | "Yes, you're right! I need to get you more info here. Either there's something going with the name i(...TRUNCATED) | "## Describe the bug\r\n\r\nGetting the following error when calling `load_dataset(\"gar1t/test\")`:(...TRUNCATED) | 67 | "\"File name too long\" error for file locks \n ## Describe the bug\r\n\r\nGetting the following err(...TRUNCATED) | [0.04933885484933853,0.0920017659664154,-0.07149408012628555,0.40184855461120605,0.4151674211025238,(...TRUNCATED) |
End of preview. Expand in Data Studio
README.md exists but content is empty.
- Downloads last month
- 12