title
stringlengths
1
290
body
stringlengths
0
228k
βŒ€
html_url
stringlengths
46
51
comments
list
pull_request
dict
number
int64
1
5.59k
is_pull_request
bool
2 classes
Support streaming allocine dataset
Support streaming allocine dataset. Fix #4562.
https://github.com/huggingface/datasets/pull/4563
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4563", "html_url": "https://github.com/huggingface/datasets/pull/4563", "diff_url": "https://github.com/huggingface/datasets/pull/4563.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4563.patch", "merged_at": "2022-06-24T16:44...
4,563
true
Dataset Viewer issue for allocine
### Link https://huggingface.co/datasets/allocine ### Description Not sure if this is a problem with `bz2` compression, but I thought these datasets could be streamed: ``` Status code: 400 Exception: AttributeError Message: 'TarContainedFile' object has no attribute 'readable' ``` ### Owner No
https://github.com/huggingface/datasets/issues/4562
[ "I removed my assignment as @huggingface/datasets should be able to answer better than me\r\n", "Let me have a look...", "Thanks for the quick fix @albertvillanova ", "Note that the underlying issue is that datasets containing TAR files are not streamable out of the box: they need being iterated with `dl_mana...
null
4,562
false
Add evaluation data to acronym_identification
null
https://github.com/huggingface/datasets/pull/4561
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4561", "html_url": "https://github.com/huggingface/datasets/pull/4561", "diff_url": "https://github.com/huggingface/datasets/pull/4561.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4561.patch", "merged_at": "2022-06-27T08:49...
4,561
true
Add evaluation metadata to imagenet-1k
null
https://github.com/huggingface/datasets/pull/4560
[ "_The documentation is not available anymore as the PR was closed or merged._", "As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4560", "html_url": "https://github.com/huggingface/datasets/pull/4560", "diff_url": "https://github.com/huggingface/datasets/pull/4560.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4560.patch", "merged_at": null }
4,560
true
Add action names in schema_guided_dstc8 dataset card
As aseked in https://huggingface.co/datasets/schema_guided_dstc8/discussions/1, I added the action names in the dataset card
https://github.com/huggingface/datasets/pull/4559
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4559", "html_url": "https://github.com/huggingface/datasets/pull/4559", "diff_url": "https://github.com/huggingface/datasets/pull/4559.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4559.patch", "merged_at": "2022-06-24T10:43...
4,559
true
Add evaluation metadata to wmt14
null
https://github.com/huggingface/datasets/pull/4558
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4558). All of your documentation changes will be reflected on that endpoint.", "As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4558", "html_url": "https://github.com/huggingface/datasets/pull/4558", "diff_url": "https://github.com/huggingface/datasets/pull/4558.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4558.patch", "merged_at": null }
4,558
true
Add evaluation metadata to wmt16
Just to confirm: we should add this metadata via GitHub and not Hub PRs for canonical datasets right?
https://github.com/huggingface/datasets/pull/4557
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4557). All of your documentation changes will be reflected on that endpoint.", "> Just to confirm: we should add this metadata via GitHub and not Hub PRs for canonical datasets right?\r\n\r\nyes :)", "As discussed with @lewtu...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4557", "html_url": "https://github.com/huggingface/datasets/pull/4557", "diff_url": "https://github.com/huggingface/datasets/pull/4557.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4557.patch", "merged_at": null }
4,557
true
Dataset Viewer issue for conll2003
### Link https://huggingface.co/datasets/conll2003/viewer/conll2003/test ### Description Seems like a cache problem with this config / split: ``` Server error Status code: 400 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: '/cache/modules/datasets_modules/datasets/conll...
https://github.com/huggingface/datasets/issues/4556
[ "Fixed, thanks." ]
null
4,556
false
Dataset Viewer issue for xtreme
### Link https://huggingface.co/datasets/xtreme/viewer/PAN-X.de/test ### Description There seems to be a problem with the cache of this config / split: ``` Server error Status code: 400 Exception: FileNotFoundError Message: [Errno 2] No such file or directory: '/cache/modules/datasets_modules/data...
https://github.com/huggingface/datasets/issues/4555
[ "Fixed, thanks." ]
null
4,555
false
Fix WMT dataset loading issue and docs update (Re-opened)
This PR is a fix for #4354 Changes are made for `wmt14`, `wmt15`, `wmt16`, `wmt17`, `wmt18`, `wmt19` and `wmt_t2t`. And READMEs are updated for the corresponding datasets. Let me know, if any additional changes are required. Thanks
https://github.com/huggingface/datasets/pull/4554
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4554", "html_url": "https://github.com/huggingface/datasets/pull/4554", "diff_url": "https://github.com/huggingface/datasets/pull/4554.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4554.patch", "merged_at": "2022-07-08T15:27...
4,554
true
Stop dropping columns in to_tf_dataset() before we load batches
`to_tf_dataset()` dropped unnecessary columns before loading batches from the dataset, but this is causing problems when using a transform, because the dropped columns might be needed to compute the transform. Since there's no real way to check which columns the transform might need, we skip dropping columns and instea...
https://github.com/huggingface/datasets/pull/4553
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq Rebasing fixed the test failures, so this should be ready to review now! There's still a failure on Win but it seems unrelated.", "Gentle ping @lhoestq ! This is a simple fix (dropping columns after loading a batch from th...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4553", "html_url": "https://github.com/huggingface/datasets/pull/4553", "diff_url": "https://github.com/huggingface/datasets/pull/4553.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4553.patch", "merged_at": "2022-07-04T18:49...
4,553
true
Tell users to upload on the hub directly
As noted in https://github.com/huggingface/datasets/pull/4534, it is still not clear that it is recommended to add datasets on the Hugging Face Hub directly instead of GitHub, so I updated some docs. Moreover since users won't be able to get reviews from us on the Hub, I added a paragraph to tell users that they can...
https://github.com/huggingface/datasets/pull/4552
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks ! I updated the two remaining files" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4552", "html_url": "https://github.com/huggingface/datasets/pull/4552", "diff_url": "https://github.com/huggingface/datasets/pull/4552.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4552.patch", "merged_at": "2022-06-26T15:39...
4,552
true
Perform hidden file check on relative data file path
Fix #4549
https://github.com/huggingface/datasets/pull/4551
[ "_The documentation is not available anymore as the PR was closed or merged._", "I'm aware of this behavior, which is tricky to solve due to fsspec's hidden file handling (see https://github.com/huggingface/datasets/issues/4115#issuecomment-1108819538). I've tested some regex patterns to address this, and they se...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4551", "html_url": "https://github.com/huggingface/datasets/pull/4551", "diff_url": "https://github.com/huggingface/datasets/pull/4551.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4551.patch", "merged_at": "2022-06-30T14:38...
4,551
true
imdb source error
## Describe the bug imdb dataset not loading ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("imdb") ``` ## Expected results ## Actual results ```bash 06/23/2022 14:45:18 - INFO - datasets.builder - Dataset not on Hf google storage. Downloading and pr...
https://github.com/huggingface/datasets/issues/4550
[ "Thanks for reporting, @Muhtasham.\r\n\r\nIndeed IMDB dataset is not accessible from yesterday, because the data is hosted on the data owners servers at Stanford (http://ai.stanford.edu/) and these are down due to a power outage originated by a fire: https://twitter.com/StanfordAILab/status/1539472302399623170?s=20...
null
4,550
false
FileNotFoundError when passing a data_file inside a directory starting with double underscores
Bug experienced in the `accelerate` CI: https://github.com/huggingface/accelerate/runs/7016055148?check_suite_focus=true This is related to https://github.com/huggingface/datasets/pull/4505 and the changes from https://github.com/huggingface/datasets/pull/4412
https://github.com/huggingface/datasets/issues/4549
[ "I have consistently experienced this bug on GitHub actions when bumping to `2.3.2`", "We're working on a fix ;)" ]
null
4,549
false
Metadata.jsonl for Imagefolder is ignored if it's in a parent directory to the splits directories/do not have "{split}_" prefix
If data contains a single `metadata.jsonl` file for several splits, it won't be included in a dataset's `data_files` and therefore ignored. This happens when a directory is structured like as follows: ``` train/ file_1.jpg file_2.jpg test/ file_3.jpg file_4.jpg metadata.jsonl ``` or like as follows:...
https://github.com/huggingface/datasets/issues/4548
[ "I agree it would be nice to support this. It doesn't fit really well in the current data_files.py, where files of each splits are separated in different folder though, maybe we have to modify a bit the logic here. \r\n\r\nOne idea would be to extend `get_patterns_in_dataset_repository` and `get_patterns_locally` t...
null
4,548
false
[CI] Fix some warnings
There are some warnings in the CI that are annoying, I tried to remove most of them
https://github.com/huggingface/datasets/pull/4547
[ "_The documentation is not available anymore as the PR was closed or merged._", "There is a CI failure only related to the missing content of the universal_dependencies dataset card, we can ignore this failure in this PR", "good catch, I thought I resolved them all sorry", "Alright it should be good now" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4547", "html_url": "https://github.com/huggingface/datasets/pull/4547", "diff_url": "https://github.com/huggingface/datasets/pull/4547.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4547.patch", "merged_at": "2022-06-28T13:59...
4,547
true
[CI] fixing seqeval install in ci by pinning setuptools-scm
The latest setuptools-scm version supported on 3.6 is 6.4.2. However for some reason circleci has version 7, which doesn't work. I fixed this by pinning the version of setuptools-scm in the circleci job Fix https://github.com/huggingface/datasets/issues/4544
https://github.com/huggingface/datasets/pull/4546
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4546", "html_url": "https://github.com/huggingface/datasets/pull/4546", "diff_url": "https://github.com/huggingface/datasets/pull/4546.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4546.patch", "merged_at": "2022-06-23T10:13...
4,546
true
Make DuplicateKeysError more user friendly [For Issue #2556]
# What does this PR do? ## Summary *DuplicateKeysError error does not provide any information regarding the examples which have the same the key.* *This information is very helpful for debugging the dataset generator script.* ## Additions - ## Changes - Changed `DuplicateKeysError Class` in `src/datase...
https://github.com/huggingface/datasets/pull/4545
[ "> Nice thanks !\r\n> \r\n> After your changes feel free to mark this PR as \"ready for review\" ;)\r\n\r\nMarking PR ready for review.\r\n\r\n@lhoestq Let me know if there is anything else required or if we are good to go ahead and merge.", "_The documentation is not available anymore as the PR was closed or mer...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4545", "html_url": "https://github.com/huggingface/datasets/pull/4545", "diff_url": "https://github.com/huggingface/datasets/pull/4545.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4545.patch", "merged_at": "2022-06-28T09:26...
4,545
true
[CI] seqeval installation fails sometimes on python 3.6
The CI sometimes fails to install seqeval, which cause the `seqeval` metric tests to fail. The installation fails because of this error: ``` Collecting seqeval Downloading seqeval-1.2.2.tar.gz (43 kB) |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Œ | 10 kB 42.1 MB/s eta 0:00:01 |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ ...
https://github.com/huggingface/datasets/issues/4544
[]
null
4,544
false
[CI] Fix upstream hub test url
Some tests were still using moon-stagign instead of hub-ci. I also updated the token to use one dedicated to `datasets`
https://github.com/huggingface/datasets/pull/4543
[ "_The documentation is not available anymore as the PR was closed or merged._", "Remaining CI failures are unrelated to this fix, merging" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4543", "html_url": "https://github.com/huggingface/datasets/pull/4543", "diff_url": "https://github.com/huggingface/datasets/pull/4543.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4543.patch", "merged_at": "2022-06-22T16:27...
4,543
true
[to_tf_dataset] Use Feather for better compatibility with TensorFlow ?
To have better performance in TensorFlow, it is important to provide lists of data files in supported formats. For example sharded TFRecords datasets are extremely performant. This is because tf.data can better leverage parallelism in this case, and load one file at a time in memory. It seems that using `tensorflow_...
https://github.com/huggingface/datasets/issues/4542
[ "This has so much potential to be great! Also I think you tagged some poor random dude on the internet whose name is also Joao, lol, edited that for you! ", "cc @sayakpaul here too, since he was interested in our new approaches to converting datasets!", "Noted and I will look into the thread in detail tomorrow ...
null
4,542
false
Fix timestamp conversion from Pandas to Python datetime in streaming mode
Arrow accepts both pd.Timestamp and datetime.datetime objects to create timestamp arrays. However a timestamp array is always converted to datetime.datetime objects. This created an inconsistency between streaming in non-streaming. e.g. the `ett` dataset outputs datetime.datetime objects in non-streaming but pd.tim...
https://github.com/huggingface/datasets/pull/4541
[ "_The documentation is not available anymore as the PR was closed or merged._", "CI failures are unrelated to this PR, merging" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4541", "html_url": "https://github.com/huggingface/datasets/pull/4541", "diff_url": "https://github.com/huggingface/datasets/pull/4541.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4541.patch", "merged_at": "2022-06-22T16:29...
4,541
true
Avoid splitting by` .py` for the file.
https://github.com/huggingface/datasets/blob/90b3a98065556fc66380cafd780af9b1814b9426/src/datasets/load.py#L272 Hello, Thanks you for this library . I was using it and I had one edge case. my home folder name ends with `.py` it is `/home/espoir.py` so anytime I am running the code to load a local module thi...
https://github.com/huggingface/datasets/issues/4540
[ "Hi @espoirMur, thanks for reporting.\r\n\r\nYou are right: that code line could be improved and made more generically valid.\r\n\r\nOn the other hand, I would suggest using `os.path.splitext` instead.\r\n\r\nAre you willing to open a PR? :)", "I will have a look.. \r\n\r\nThis weekend .. ", "@albertvillanova ...
null
4,540
false
Replace deprecated logging.warn with logging.warning
Replace `logging.warn` (deprecated in [Python 2.7, 2011](https://github.com/python/cpython/commit/04d5bc00a219860c69ea17eaa633d3ab9917409f)) with `logging.warning` (added in [Python 2.3, 2003](https://github.com/python/cpython/commit/6fa635df7aa88ae9fd8b41ae42743341316c90f7)). * https://docs.python.org/3/library/log...
https://github.com/huggingface/datasets/pull/4539
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4539", "html_url": "https://github.com/huggingface/datasets/pull/4539", "diff_url": "https://github.com/huggingface/datasets/pull/4539.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4539.patch", "merged_at": "2022-06-22T12:51...
4,539
true
Dataset Viewer issue for Pile of Law
### Link https://huggingface.co/datasets/pile-of-law/pile-of-law ### Description Hi, I would like to turn off the dataset viewer for our dataset without enabling access requests. To comply with upstream dataset creator requests/licenses, we would like to make sure that the data is not indexed by search engines...
https://github.com/huggingface/datasets/issues/4538
[ "Hi @Breakend, yes – we'll propose a solution today", "Thanks so much, I appreciate it!", "Thanks so much for adding the docs. I was able to successfully hide the viewer using the \r\n```\r\nviewer: false\r\n```\r\nflag in the README.md of the dataset. I'm closing the issue because this is resolved. Thanks agai...
null
4,538
false
Fix WMT dataset loading issue and docs update
This PR is a fix for #4354 Changes are made for `wmt14`, `wmt15`, `wmt16`, `wmt17`, `wmt18`, `wmt19` and `wmt_t2t`. And READMEs are updated for the corresponding datasets. As I am on a M1 Mac, I am not able to create a virtual `dev` environment using `pip install -e ".[dev]"`. Issue is with `tensorflow-text` not...
https://github.com/huggingface/datasets/pull/4537
[ "The PR branch now has some commits unrelated to the changes, probably due to rebasing. Can you please close this PR and open a new one from a new branch? You can use `git cherry-pick` to preserve the relevant changes:\r\n```bash\r\ngit checkout master\r\ngit remote add upstream git@github.com:huggingface/datasets...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4537", "html_url": "https://github.com/huggingface/datasets/pull/4537", "diff_url": "https://github.com/huggingface/datasets/pull/4537.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4537.patch", "merged_at": null }
4,537
true
Properly raise FileNotFound even if the dataset is private
`tests/test_load.py::test_load_streaming_private_dataset` was failing because the hub now returns 401 when getting the HfApi.dataset_info of a dataset without authentication. `load_dataset` was raising ConnectionError, while it should be FileNoteFoundError since it first checks for local files before checking the Hub. ...
https://github.com/huggingface/datasets/pull/4536
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4536", "html_url": "https://github.com/huggingface/datasets/pull/4536", "diff_url": "https://github.com/huggingface/datasets/pull/4536.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4536.patch", "merged_at": "2022-06-28T10:36...
4,536
true
Add `batch_size` parameter when calling `add_faiss_index` and `add_faiss_index_from_external_arrays`
Currently, even though the `batch_size` when adding vectors to the FAISS index can be tweaked in `FaissIndex.add_vectors()`, the function `ArrowDataset.add_faiss_index` doesn't have either the parameter `batch_size` to be propagated to the nested `FaissIndex.add_vectors` function or `*args, **kwargs`, so on, this PR ad...
https://github.com/huggingface/datasets/pull/4535
[ "Also, I had a doubt while checking the code related to the indices... \r\n\r\n@lhoestq, there's a value in `config.py` named `DATASET_INDICES_FILENAME` which has the arrow extension (which I assume it should be `indices.faiss`, as the Elastic Search indices are not stored in a file, but not sure), and it's just us...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4535", "html_url": "https://github.com/huggingface/datasets/pull/4535", "diff_url": "https://github.com/huggingface/datasets/pull/4535.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4535.patch", "merged_at": "2022-06-27T16:14...
4,535
true
Add `tldr_news` dataset
This PR aims at adding support for a news dataset: `tldr news`. This dataset is based on the daily [tldr tech newsletter](https://tldr.tech/newsletter) and contains a `headline` as well as a `content` for every piece of news contained in a newsletter.
https://github.com/huggingface/datasets/pull/4534
[ "Hey @lhoestq, \r\nSorry for opening a PR, I was following the guide [here](https://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md)! Thanks for the review anyway, I will follow the instructions you sent πŸ˜ƒ ", "Thanks, we will update the guide ;)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4534", "html_url": "https://github.com/huggingface/datasets/pull/4534", "diff_url": "https://github.com/huggingface/datasets/pull/4534.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4534.patch", "merged_at": null }
4,534
true
Timestamp not returned as datetime objects in streaming mode
As reported in (internal) https://github.com/huggingface/datasets-server/issues/397 ```python >>> from datasets import load_dataset >>> dataset = load_dataset("ett", name="h2", split="test", streaming=True) >>> d = next(iter(dataset)) >>> d['start'] Timestamp('2016-07-01 00:00:00') ``` while loading in non-...
https://github.com/huggingface/datasets/issues/4533
[]
null
4,533
false
Add Video feature
The following adds a `Video` feature for encoding/decoding videos on the fly from in memory bytes. It uses my own `encoded-video` library which is basically `pytorchvideo`'s encoded video but with all the `torch` specific stuff stripped out. Because of that, and because the tool I used under the hood is not very mature...
https://github.com/huggingface/datasets/pull/4532
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4532). All of your documentation changes will be reflected on that endpoint.", "@nateraw do you have any plans to continue this pr? Or should I write a custom loader script to use my video dataset in the hub?", "@fcakyon I th...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4532", "html_url": "https://github.com/huggingface/datasets/pull/4532", "diff_url": "https://github.com/huggingface/datasets/pull/4532.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4532.patch", "merged_at": null }
4,532
true
Dataset Viewer issue for CSV datasets
### Link https://huggingface.co/datasets/scikit-learn/breast-cancer-wisconsin ### Description I'm populating CSV datasets [here](https://huggingface.co/scikit-learn) but the viewer is not enabled and it looks for a dataset loading script, the datasets aren't on queue as well. You can replicate the problem by sim...
https://github.com/huggingface/datasets/issues/4531
[ "this should now be fixed", "Confirmed, it's fixed now. Thanks for reporting, and thanks @coyotte508 for fixing it\r\n\r\n<img width=\"1123\" alt=\"Capture d’écran 2022-06-21 aΜ€ 10 28 05\" src=\"https://user-images.githubusercontent.com/1676121/174753833-1b453a5a-6a90-4717-bca1-1b5fc6b75e4a.png\">\r\n" ]
null
4,531
false
Add AudioFolder packaged loader
will close #3964 AudioFolder is almost identical to ImageFolder except for inferring labels is not the default behavior (`drop_labels` is set to True in config), the option of inferring them is preserved though. The weird thing is happening with the `test_data_files_with_metadata_and_archives` when `streaming` i...
https://github.com/huggingface/datasets/pull/4530
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq @mariosasko I don't know what to do with the test, do you have any ideas? :)", "also it's passed in `pyarrow_latest_WIN`", "If the error only happens on 3.6, maybe #4460 can help ^^' It seems to work in 3.7 on the window...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4530", "html_url": "https://github.com/huggingface/datasets/pull/4530", "diff_url": "https://github.com/huggingface/datasets/pull/4530.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4530.patch", "merged_at": "2022-08-22T14:20...
4,530
true
Ecoset
## Adding a Dataset - **Name:** *Ecoset* - **Description:** *https://www.kietzmannlab.org/ecoset/* - **Paper:** *https://doi.org/10.1073/pnas.2011417118* - **Data:** *https://codeocean.com/capsule/9570390/tree/v1* - **Motivation:** **Ecoset** was created as a clean and ecologically valid alternative to **Imagen...
https://github.com/huggingface/datasets/issues/4529
[ "Hi! Very cool dataset! I answered your questions on the forum. Also, feel free to comment `#self-assign` on this issue to self-assign it." ]
null
4,529
false
Memory leak when iterating a Dataset
e## Describe the bug It seems that memory never gets freed after iterating a `Dataset` (using `.map()` or a simple `for` loop) ## Steps to reproduce the bug ```python import gc import logging import time import pyarrow from datasets import load_dataset from tqdm import trange import os, psutil logging.ba...
https://github.com/huggingface/datasets/issues/4528
[ "Is someone assigned to this issue?", "The same issue is being debugged here: https://github.com/huggingface/datasets/issues/4883\r\n", "Here is a modified repro example that makes it easier to see the leak:\r\n\r\n```\r\n$ cat ds2.py\r\nimport gc, sys\r\nimport time\r\nfrom datasets import load_dataset\r\nimpo...
null
4,528
false
Dataset Viewer issue for vadis/sv-ident
### Link https://huggingface.co/datasets/vadis/sv-ident ### Description The dataset preview does not work: ``` Server Error Status code: 400 Exception: Status400Error Message: The dataset does not exist. ``` However, the dataset is streamable and works locally: ```python In [1]: from dataset...
https://github.com/huggingface/datasets/issues/4527
[ "Fixed, thanks!\r\n![Uploading Capture d’écran 2022-06-21 aΜ€ 18.42.40.png…]()\r\n\r\n" ]
null
4,527
false
split cache used when processing different split
## Describe the bug` ``` ds1 = load_dataset('squad', split='validation') ds2 = load_dataset('squad', split='train') ds1 = ds1.map(some_function) ds2 = ds2.map(some_function) assert ds1 == ds2 ``` This happens when ds1 and ds2 are created in `pytorch_lightning.DataModule` through ``` class myDataModule: ...
https://github.com/huggingface/datasets/issues/4526
[ "I was not able to reproduce this behavior (I tried without using pytorch lightning though, since I don't know what code you ran in pytorch lightning to get this).\r\n\r\nIf you can provide a MWE that would be perfect ! :)", "Hi, I think the issue happened because I was loading datasets under an `if` ... `else` s...
null
4,526
false
Out of memory error on workers while running Beam+Dataflow
## Describe the bug While running the preprocessing of the natural_question dataset (see PR #4368), there is an issue for the "default" config (train+dev files). Previously we ran the preprocessing for the "dev" config (only dev files) with success. Train data files are larger than dev ones and apparently worker...
https://github.com/huggingface/datasets/issues/4525
[ "Some naive ideas to cope with this:\r\n- enable more RAM on each worker\r\n- force the spanning of more workers\r\n- others?", "@albertvillanova We were finally able to process the full NQ dataset on our machines using 600 gb with 5 workers. Maybe these numbers will work for you as well.", "Thanks a lot for th...
null
4,525
false
Downloading via Apache Pipeline, client cancelled (org.apache.beam.vendor.grpc.v1p43p2.io.grpc.StatusRuntimeException)
## Describe the bug When downloading some `wikipedia` languages (in particular, I'm having a hard time with Spanish, Cebuano, and Russian) via FlinkRunner, I encounter the exception in the title. I have been playing with package versions a lot, because unfortunately, the different dependencies required by these packag...
https://github.com/huggingface/datasets/issues/4524
[ "Hi @dan-the-meme-man, thanks for reporting.\r\n\r\nWe are investigating a similar issue but with Beam+Dataflow (instead of Beam+Flink): \r\n- #4525\r\n\r\nIn order to go deeper into the root cause, we need as much information as possible: logs from the main process + logs from the workers are very informative.\r\n...
null
4,524
false
Update download url and improve card of `cats_vs_dogs` dataset
Improve the download URL (reported here: https://huggingface.co/datasets/cats_vs_dogs/discussions/1), remove the `image_file_path` column (not used in Transformers, so it should be safe) and add more info to the card.
https://github.com/huggingface/datasets/pull/4523
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4523", "html_url": "https://github.com/huggingface/datasets/pull/4523", "diff_url": "https://github.com/huggingface/datasets/pull/4523.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4523.patch", "merged_at": "2022-06-21T14:13...
4,523
true
Try to reduce the number of datasets that require manual download
> Currently, 41 canonical datasets require manual download. I checked their scripts and I'm pretty sure this number can be reduced to β‰ˆ 30 by not relying on bash scripts to download data, hosting data directly on the Hub when the license permits, etc. Then, we will mostly be left with datasets with restricted access, w...
https://github.com/huggingface/datasets/issues/4522
[]
null
4,522
false
Datasets method `.map` not hashing
## Describe the bug Datasets method `.map` not hashing, even with an empty no-op function ## Steps to reproduce the bug ```python from datasets import load_dataset # download 9MB dummy dataset ds = load_dataset("hf-internal-testing/librispeech_asr_dummy", "clean") def prepare_dataset(batch): return(b...
https://github.com/huggingface/datasets/issues/4521
[ "Fix posted: https://github.com/huggingface/datasets/issues/4506#issuecomment-1157417219", "Didn't realize it's a bug when I asked the question yesterday! Free free to post an answer if you are sure the cause has been addressed.\r\n\r\nhttps://stackoverflow.com/questions/72664827/can-pickle-dill-foo-but-not-lambd...
null
4,521
false
Failure to hash `dataclasses` - results in functions that cannot be hashed or cached in `.map`
Dataclasses cannot be hashed. As a result, they cannot be hashed or cached if used in the `.map` method. Dataclasses are used extensively in Transformers examples scripts: (c.f. [CTC example](https://github.com/huggingface/transformers/blob/main/examples/pytorch/speech-recognition/run_speech_recognition_ctc.py)). Since...
https://github.com/huggingface/datasets/issues/4520
[ "I think this has been fixed by #4516, let me know if you encounter this again :)\r\n\r\nI re-ran your code in 3.7 and 3.9 and it works fine", "Thank you!" ]
null
4,520
false
Create new sections for audio and vision in guides
This PR creates separate sections in the guides for audio, vision, text, and general usage so it is easier for users to find loading, processing, or sharing guides specific to the dataset type they're working with. It'll also allow us to scale the docs to additional dataset types - like time series, tabular, etc. - whi...
https://github.com/huggingface/datasets/pull/4519
[ "_The documentation is not available anymore as the PR was closed or merged._", "Ready for review!\r\n\r\nThe `toctree` is a bit longer now with the sections. I think if we keep the audio/vision/text/dataset repository sections collapsed by default, and keep the general usage expanded, it may look a little cleane...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4519", "html_url": "https://github.com/huggingface/datasets/pull/4519", "diff_url": "https://github.com/huggingface/datasets/pull/4519.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4519.patch", "merged_at": "2022-07-07T15:24...
4,519
true
Patch tests for hfh v0.8.0
This PR patches testing utilities that would otherwise fail with hfh v0.8.0.
https://github.com/huggingface/datasets/pull/4518
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4518", "html_url": "https://github.com/huggingface/datasets/pull/4518", "diff_url": "https://github.com/huggingface/datasets/pull/4518.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4518.patch", "merged_at": "2022-06-17T16:06...
4,518
true
Add tags for task_ids:summarization-* and task_categories:summarization*
yaml header at top of README.md file was edited to add task tags because I couldn't find the existing tags in the json separate Pull Request will modify dataset_infos.json to add these tags The Enron dataset (dataset id aeslc) is only tagged with: arxiv:1906.03497' languages:en pretty_name:AESLC ...
https://github.com/huggingface/datasets/pull/4517
[ "Associated community discussion is [here](https://huggingface.co/datasets/aeslc/discussions/1).\r\nPaper referenced in the `dataset_infos.json` is [here](https://arxiv.org/pdf/1906.03497.pdf). It mentions the _email-subject-generation_ task, which is not a tag mentioned in any other dataset so it was not added in...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4517", "html_url": "https://github.com/huggingface/datasets/pull/4517", "diff_url": "https://github.com/huggingface/datasets/pull/4517.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4517.patch", "merged_at": "2022-07-08T15:02...
4,517
true
Fix hashing for python 3.9
In python 3.9, pickle hashes the `glob_ids` dictionary in addition to the `globs` of a function. Therefore the test at `tests/test_fingerprint.py::RecurseDumpTest::test_recurse_dump_for_function_with_shuffled_globals` is currently failing for python 3.9 To make hashing deterministic when the globals are not in th...
https://github.com/huggingface/datasets/pull/4516
[ "_The documentation is not available anymore as the PR was closed or merged._", "What do you think @albertvillanova ?" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4516", "html_url": "https://github.com/huggingface/datasets/pull/4516", "diff_url": "https://github.com/huggingface/datasets/pull/4516.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4516.patch", "merged_at": "2022-06-28T13:23...
4,516
true
Add uppercased versions of image file extensions for automatic module inference
Adds the uppercased versions of the image file extensions to the supported extensions. Another approach would be to call `.lower()` on extensions while resolving data files, but uppercased extensions are not something we want to encourage out of the box IMO unless they are commonly used (as they are in the vision d...
https://github.com/huggingface/datasets/pull/4515
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4515", "html_url": "https://github.com/huggingface/datasets/pull/4515", "diff_url": "https://github.com/huggingface/datasets/pull/4515.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4515.patch", "merged_at": "2022-06-16T17:11...
4,515
true
Allow .JPEG as a file extension
## Describe the bug When loading image data, HF datasets seems to recognize `.jpg` and `.jpeg` file extensions, but not e.g. .JPEG. As the naming convention .JPEG is used in important datasets such as imagenet, I would welcome if according extensions like .JPEG or .JPG would be allowed. ## Steps to reproduce the bu...
https://github.com/huggingface/datasets/issues/4514
[ "Hi, thanks for reporting! I've opened a PR with the fix.", "Wow, that was quick! Thank you very much πŸ™ " ]
null
4,514
false
Update Google Cloud Storage documentation and add Azure Blob Storage example
While I was going through the πŸ€— Datasets documentation of the Cloud storage filesystems at https://huggingface.co/docs/datasets/filesystems, I realized that the Google Cloud Storage documentation could be improved e.g. bullet point says "Load your dataset" when the actual call was to "Save your dataset", in-line code ...
https://github.com/huggingface/datasets/pull/4513
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi @stevhliu, I've kept the `>>>` before all the in-line code comments as it was done like that in the default S3 example that was already there, I assume that it's done like that just for readiness, let me know whether we should rem...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4513", "html_url": "https://github.com/huggingface/datasets/pull/4513", "diff_url": "https://github.com/huggingface/datasets/pull/4513.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4513.patch", "merged_at": "2022-06-23T16:54...
4,513
true
Add links to vision tasks scripts in ADD_NEW_DATASET template
Add links to vision dataset scripts in the ADD_NEW_DATASET template.
https://github.com/huggingface/datasets/pull/4512
[ "_The documentation is not available anymore as the PR was closed or merged._", "The CI failure is unrelated to the PR's changes. Merging." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4512", "html_url": "https://github.com/huggingface/datasets/pull/4512", "diff_url": "https://github.com/huggingface/datasets/pull/4512.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4512.patch", "merged_at": "2022-07-08T13:56...
4,512
true
Support all negative values in ClassLabel
We usually use -1 to represent a missing label, but we should also support any negative values (some users use -100 for example). This is a regression from `datasets` 2.3 Fix https://github.com/huggingface/datasets/issues/4508
https://github.com/huggingface/datasets/pull/4511
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for this fix! I'm not sure what the release timeline is, but FYI #4508 is a breaking issue for transformer token classification using Trainer and PyTorch. PyTorch defaults to -100 as the ignored label for [negative log loss](h...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4511", "html_url": "https://github.com/huggingface/datasets/pull/4511", "diff_url": "https://github.com/huggingface/datasets/pull/4511.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4511.patch", "merged_at": "2022-06-16T13:54...
4,511
true
Add regression test for `ArrowWriter.write_batch` when batch is empty
As spotted by @cccntu in #4502, there's a logic bug in `ArrowWriter.write_batch` as the if-statement to handle the empty batches as detailed in the docstrings of the function ("Ignores the batch if it appears to be empty, preventing a potential schema update of unknown types."), the current if-statement is not handling...
https://github.com/huggingface/datasets/pull/4510
[ "_The documentation is not available anymore as the PR was closed or merged._", "As mentioned by @lhoestq, the current behavior is correct and we should not expect batches with different columns, in that case, the if should fail, as the values of the batch can be empty, but not the actual `batch_examples` value."...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4510", "html_url": "https://github.com/huggingface/datasets/pull/4510", "diff_url": "https://github.com/huggingface/datasets/pull/4510.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4510.patch", "merged_at": "2022-06-16T12:28...
4,510
true
Support skipping Parquet to Arrow conversion when using Beam
null
https://github.com/huggingface/datasets/pull/4509
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4509). All of your documentation changes will be reflected on that endpoint.", "When #4724 is merged, we can just pass `file_format=\"parquet\"` to `download_and_prepare` and it will output parquet fiels without converting to a...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4509", "html_url": "https://github.com/huggingface/datasets/pull/4509", "diff_url": "https://github.com/huggingface/datasets/pull/4509.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4509.patch", "merged_at": null }
4,509
true
cast_storage method from datasets.features
## Describe the bug A bug occurs when mapping a function to a dataset object. I ran the same code with the same data yesterday and it worked just fine. It works when i run locally on an old version of datasets. ## Steps to reproduce the bug Steps are: - load whatever datset - write a preprocessing function such ...
https://github.com/huggingface/datasets/issues/4508
[ "Hi! We've recently added a check to the `ClassLabel` type to ensure the values are in the valid label range `-1, 0, ..., num_classes-1` (-1 is used for missing values). The error in your case happens only if the `labels` column is of type `Sequence(ClassLabel(...))` before the `map` call and can be avoided by call...
null
4,508
false
How to let `load_dataset` return a `Dataset` instead of `DatasetDict` in customized loading script
If the dataset does not need splits, i.e., no training and validation split, more like a table. How can I let the `load_dataset` function return a `Dataset` object directly rather than return a `DatasetDict` object with only one key-value pair. Or I can paraphrase the question in the following way: how to skip `_spl...
https://github.com/huggingface/datasets/issues/4507
[ "Hi @liyucheng09.\r\n\r\nUsers can pass the `split` parameter to `load_dataset`. For example, if your split name is \"train\",\r\n```python\r\nds = load_dataset(\"dataset_name\", split=\"train\")\r\n```\r\nwill return a `Dataset` instance.", "@albertvillanova Thanks! I can't believe I didn't know this feature til...
null
4,507
false
Failure to hash (and cache) a `.map(...)` (almost always) - using this method can produce incorrect results
## Describe the bug Sometimes I get messages about not being able to hash a method: `Parameter 'function'=<function StupidDataModule._separate_speaker_id_from_dialogue at 0x7f1b27180d30> of the transform datasets.arrow_dataset.Dataset. _map_single couldn't be hashed properly, a random hash was used instead. Make sur...
https://github.com/huggingface/datasets/issues/4506
[ "Important info:\r\n\r\nAs hashes are generated randomly for functions, it leads to **false identifying some results as already hashed** (mapping function is not executed after a method update) when there's a `pytorch_lightning.seed_everything(123)`", "@lhoestq\r\nseems like quite critical stuff for me, if I'm no...
null
4,506
false
Fix double dots in data files
As mentioned in https://github.com/huggingface/transformers/pull/17715 `data_files` can't find a file if the path contains double dots `/../`. This has been introduced in https://github.com/huggingface/datasets/pull/4412, by trying to ignore hidden files and directories (i.e. if they start with a dot) I fixed this a...
https://github.com/huggingface/datasets/pull/4505
[ "_The documentation is not available anymore as the PR was closed or merged._", "The CI fails are unrelated to this PR (apparently something related to `seqeval` on windows) - merging :)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4505", "html_url": "https://github.com/huggingface/datasets/pull/4505", "diff_url": "https://github.com/huggingface/datasets/pull/4505.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4505.patch", "merged_at": "2022-06-15T17:05...
4,505
true
Can you please add the Stanford dog dataset?
## Adding a Dataset - **Name:** *Stanford dog dataset* - **Description:** *The dataset is about 120 classes for a total of 20.580 images. You can find the dataset here http://vision.stanford.edu/aditya86/ImageNetDogs/* - **Paper:** *http://vision.stanford.edu/aditya86/ImageNetDogs/* - **Data:** *[link to the Github...
https://github.com/huggingface/datasets/issues/4504
[ "would you like to give it a try, @dgrnd4? (maybe with the help of the dataset author?)", "@julien-c i am sorry but I have no idea about how it works: can I add the dataset by myself, following \"instructions to add a new dataset\"?\r\nCan I add a dataset even if it's not mine? (it's public in the link that I wro...
null
4,504
false
Refactor and add metadata to fever dataset
Related to: #4452 and #3792.
https://github.com/huggingface/datasets/pull/4503
[ "_The documentation is not available anymore as the PR was closed or merged._", "But this is somehow fever v3 dataset (see this link https://fever.ai/ under the dropdown menu called Datasets). Our fever dataset already contains v1 and v2 configs. Then, I added this as if v3 config (but named feverous instead of v...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4503", "html_url": "https://github.com/huggingface/datasets/pull/4503", "diff_url": "https://github.com/huggingface/datasets/pull/4503.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4503.patch", "merged_at": "2022-07-06T11:41...
4,503
true
Logic bug in arrow_writer?
https://github.com/huggingface/datasets/blob/88a902d6474fae8d793542d57a4f3b0d187f3c5b/src/datasets/arrow_writer.py#L475-L488 I got some error, and I found it's caused by `batch_examples` being `{}`. I wonder if the code should be as follows: ``` - if batch_examples and len(next(iter(batch_examples.values())...
https://github.com/huggingface/datasets/issues/4502
[ "Hi @cccntu you're right, as when `batch_examples={}` the current if-statement won't be triggered as the condition won't be satisfied, I'll prepare a PR to address it as well as add the regression tests so that this issue is handled properly.", "Hi @alvarobartt ,\r\nThanks for answering. Do you know when and why ...
null
4,502
false
Corrected broken links in doc
null
https://github.com/huggingface/datasets/pull/4501
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4501", "html_url": "https://github.com/huggingface/datasets/pull/4501", "diff_url": "https://github.com/huggingface/datasets/pull/4501.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4501.patch", "merged_at": "2022-06-15T15:00...
4,501
true
Add `concatenate_datasets` for iterable datasets
`concatenate_datasets` currently only supports lists of `datasets.Dataset`, not lists of `datasets.IterableDataset` like `interleave_datasets` Fix https://github.com/huggingface/datasets/issues/2564 I also moved `_interleave_map_style_datasets` from combine.py to arrow_dataset.py, since the logic depends a lot on...
https://github.com/huggingface/datasets/pull/4500
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks ! I addressed your comments :)\r\n\r\n> There is a slight difference in concatenate_datasets between the version for map-style datasets and the one for iterable datasets\r\n\r\nIndeed, here is what I did to fix this:\r\n\r\n- ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4500", "html_url": "https://github.com/huggingface/datasets/pull/4500", "diff_url": "https://github.com/huggingface/datasets/pull/4500.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4500.patch", "merged_at": "2022-06-28T21:15...
4,500
true
fix ETT m1/m2 test/val dataset
https://huggingface.co/datasets/ett/discussions/1
https://github.com/huggingface/datasets/pull/4499
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thansk for the fix ! Can you regenerate the datasets_infos.json please ? This way it will update the expected number of examples in the test and val splits", "ah yes!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4499", "html_url": "https://github.com/huggingface/datasets/pull/4499", "diff_url": "https://github.com/huggingface/datasets/pull/4499.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4499.patch", "merged_at": "2022-06-15T14:45...
4,499
true
WER and CER > 1
## Describe the bug It seems that in some cases in which the `prediction` is longer than the `reference` we may have word/character error rate higher than 1 which is a bit odd. If it's a real bug I think I can solve it with a PR changing [this](https://github.com/huggingface/datasets/blob/master/metrics/wer/wer.py#...
https://github.com/huggingface/datasets/issues/4498
[ "WER can have values bigger than 1.0, this is expected when there are too many insertions\r\n\r\nFrom [wikipedia](https://en.wikipedia.org/wiki/Word_error_rate):\r\n> Note that since N is the number of words in the reference, the word error rate can be larger than 1.0" ]
null
4,498
false
Re-add download_manager module in utils
https://github.com/huggingface/datasets/pull/4384 moved `datasets.utils.download_manager` to `datasets.download.download_manager` This breaks `evaluate` which imports `DownloadMode` from `datasets.utils.download_manager` This PR re-adds `datasets.utils.download_manager` without circular imports. We could also...
https://github.com/huggingface/datasets/pull/4497
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for the fix.\r\n\r\nI'm wondering how this fixes backward compatibility...\r\n\r\nExecuting this code:\r\n```python\r\nfrom datasets.utils.download_manager import DownloadMode\r\n```\r\nwe will have\r\n```python\r\nDownloadMod...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4497", "html_url": "https://github.com/huggingface/datasets/pull/4497", "diff_url": "https://github.com/huggingface/datasets/pull/4497.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4497.patch", "merged_at": "2022-06-15T10:23...
4,497
true
Replace `assertEqual` with `assertTupleEqual` in unit tests for verbosity
As detailed in #4419 and as suggested by @mariosasko, we could replace the `assertEqual` assertions with `assertTupleEqual` when the assertion is between Tuples, in order to make the tests more verbose.
https://github.com/huggingface/datasets/pull/4496
[ "_The documentation is not available anymore as the PR was closed or merged._", "FYI I used the following regex to look for the `assertEqual` statements where the assertion was being done over a Tuple: `self.assertEqual(.*, \\(.*,)(\\)\\))$`, hope this is useful!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4496", "html_url": "https://github.com/huggingface/datasets/pull/4496", "diff_url": "https://github.com/huggingface/datasets/pull/4496.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4496.patch", "merged_at": "2022-07-07T16:55...
4,496
true
Fix patching module that doesn't exist
Reported in https://github.com/huggingface/huggingface_hub/runs/6894703718?check_suite_focus=true When trying to patch `scipy.io.loadmat`: ```python ModuleNotFoundError: No module named 'scipy' ``` Instead it shouldn't raise an error and do nothing Bug introduced by #4375 Fix https://github.com/hugging...
https://github.com/huggingface/datasets/pull/4495
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4495", "html_url": "https://github.com/huggingface/datasets/pull/4495", "diff_url": "https://github.com/huggingface/datasets/pull/4495.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4495.patch", "merged_at": "2022-06-15T08:54...
4,495
true
Patching fails for modules that are not installed or don't exist
Reported in https://github.com/huggingface/huggingface_hub/runs/6894703718?check_suite_focus=true When trying to patch `scipy.io.loadmat`: ```python ModuleNotFoundError: No module named 'scipy' ``` Instead it shouldn't raise an error and do nothing We use patching to extend such functions to support remot...
https://github.com/huggingface/datasets/issues/4494
[]
null
4,494
false
Add `@transmit_format` in `flatten`
As suggested by @mariosasko in https://github.com/huggingface/datasets/pull/4411, we should include the `@transmit_format` decorator to `flatten`, `rename_column`, and `rename_columns` so as to ensure that the value of `_format_columns` in an `ArrowDataset` is properly updated. **Edit**: according to @mariosasko com...
https://github.com/huggingface/datasets/pull/4493
[ "@mariosasko please let me know whether we need to include some sort of tests to make sure that the decorator is working as expected. Thanks! πŸ€— ", "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4493). All of your documentation changes will be reflected on that endpoint.", ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4493", "html_url": "https://github.com/huggingface/datasets/pull/4493", "diff_url": "https://github.com/huggingface/datasets/pull/4493.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4493.patch", "merged_at": null }
4,493
true
Pin the revision in imagenet download links
Use the commit sha in the data files URLs of the imagenet-1k download script, in case we want to restructure the data files in the future. For example we may split it into many more shards for better paralellism. cc @mariosasko
https://github.com/huggingface/datasets/pull/4492
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4492", "html_url": "https://github.com/huggingface/datasets/pull/4492", "diff_url": "https://github.com/huggingface/datasets/pull/4492.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4492.patch", "merged_at": "2022-06-14T17:25...
4,492
true
Dataset Viewer issue for Pavithree/test
### Link https://huggingface.co/datasets/Pavithree/test ### Description I have extracted the subset of original eli5 dataset found at hugging face. However, while loading the dataset It throws ArrowNotImplementedError: Unsupported cast from string to null using function cast_null error. Is there anything missi...
https://github.com/huggingface/datasets/issues/4491
[ "This issue can be resolved according to this post https://stackoverflow.com/questions/70566660/parquet-with-null-columns-on-pyarrow. It looks like first data entry in the json file must not have any null values as pyarrow uses this first file to infer schema for entire dataset." ]
null
4,491
false
Use `torch.nested_tensor` for arrays of varying length in torch formatter
Use `torch.nested_tensor` for arrays of varying length in `TorchFormatter`. The PyTorch API of nested tensors is in the prototype stage, so wait for it to become more mature.
https://github.com/huggingface/datasets/issues/4490
[]
null
4,490
false
Add SV-Ident dataset
null
https://github.com/huggingface/datasets/pull/4489
[ "Hi @e-tornike, thanks a lot for adding this interesting dataset.\r\n\r\nRecently at Hugging Face, we have decided to give priority to adding datasets directly on the Hub. Would you mind to transfer your loading script to the Hub? You could create a dedicated org namespace, so that your dataset would be accessible ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4489", "html_url": "https://github.com/huggingface/datasets/pull/4489", "diff_url": "https://github.com/huggingface/datasets/pull/4489.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4489.patch", "merged_at": null }
4,489
true
Update PASS dataset version
Update the PASS dataset to version v3 (the newest one) from the [version history](https://github.com/yukimasano/PASS/blob/main/version_history.txt). PS: The older versions are not exposed as configs in the script because v1 was removed from Zenodo, and the same thing will probably happen to v2.
https://github.com/huggingface/datasets/pull/4488
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4488", "html_url": "https://github.com/huggingface/datasets/pull/4488", "diff_url": "https://github.com/huggingface/datasets/pull/4488.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4488.patch", "merged_at": "2022-06-14T16:32...
4,488
true
Support streaming UDHR dataset
This PR: - Adds support for streaming UDHR dataset - Adds the BCP 47 language code as feature
https://github.com/huggingface/datasets/pull/4487
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4487", "html_url": "https://github.com/huggingface/datasets/pull/4487", "diff_url": "https://github.com/huggingface/datasets/pull/4487.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4487.patch", "merged_at": "2022-06-15T04:59...
4,487
true
Add CCAgT dataset
As described in #4075 I could not generate the dummy data. Also, on the data repository isn't provided the split IDs, but I copy the functions that provide the correct data split. In summary, to have a better distribution, the data in this dataset should be separated based on the amount of NORs in each image.
https://github.com/huggingface/datasets/pull/4486
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi! Excellent job @johnnv1! There were typos/missing words in the card, so I took the liberty to rewrite some parts to make them easier to understand. Let me know if you are ok with the changes. Also, feel free to add some info under...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4486", "html_url": "https://github.com/huggingface/datasets/pull/4486", "diff_url": "https://github.com/huggingface/datasets/pull/4486.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4486.patch", "merged_at": null }
4,486
true
Fix cast to null
It currently fails with `ArrowNotImplementedError` instead of `TypeError` when one tries to cast integer to null type. Because if this, type inference breaks when one replaces null values with integers in `map` (it first tries to cast to the previous type before inferring the new type). Fix https://github.com/hug...
https://github.com/huggingface/datasets/pull/4485
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4485", "html_url": "https://github.com/huggingface/datasets/pull/4485", "diff_url": "https://github.com/huggingface/datasets/pull/4485.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4485.patch", "merged_at": "2022-06-14T13:34...
4,485
true
Better ImportError message when a dataset script dependency is missing
When a depenency is missing for a dataset script, an ImportError message is shown, with a tip to install the missing dependencies. This message is not ideal at the moment: it may show duplicate dependencies, and is not very readable. I improved it from ``` ImportError: To be able to use bigbench, you need to insta...
https://github.com/huggingface/datasets/pull/4484
[ "_The documentation is not available anymore as the PR was closed or merged._", "Discussed offline with @mariosasko, merging :)", "Fwiw, i think this same issue is occurring on the datasets website page, where preview isn't available due to the `bigbench` import error", "For the preview of BigBench datasets, ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4484", "html_url": "https://github.com/huggingface/datasets/pull/4484", "diff_url": "https://github.com/huggingface/datasets/pull/4484.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4484.patch", "merged_at": "2022-06-13T13:50...
4,484
true
Dataset.map throws pyarrow.lib.ArrowNotImplementedError when converting from list of empty lists
## Describe the bug Dataset.map throws pyarrow.lib.ArrowNotImplementedError: Unsupported cast from int64 to null using function cast_null when converting from a type of 'empty lists' to 'lists with some type'. This appears to be due to the interaction of arrow internals and some assumptions made by datasets. T...
https://github.com/huggingface/datasets/issues/4483
[ "Hi @sanderland ! Thanks for reporting :) This is a bug, I opened a PR to fix it. We'll do a new release soon\r\n\r\nIn the meantime you can fix it by specifying in advance that the \"label\" are integers:\r\n```python\r\nimport numpy as np\r\n\r\nds = Dataset.from_dict(\r\n {\r\n \"text\": [\"the lazy do...
null
4,483
false
Test that TensorFlow is not imported on startup
TF takes some time to be imported, and also uses some GPU memory. I just added a test to make sure that in the future it's never imported by default when ```python import datasets ``` is called. Right now this fails because `huggingface_hub` does import tensorflow (though this is fixed now on their `main` bra...
https://github.com/huggingface/datasets/pull/4482
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4482). All of your documentation changes will be reflected on that endpoint." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4482", "html_url": "https://github.com/huggingface/datasets/pull/4482", "diff_url": "https://github.com/huggingface/datasets/pull/4482.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4482.patch", "merged_at": null }
4,482
true
Fix iwslt2017
The files were moved to google drive, I hosted them on the Hub instead (ok according to the license) I also updated the `datasets_infos.json`
https://github.com/huggingface/datasets/pull/4481
[ "_The documentation is not available anymore as the PR was closed or merged._", "CI fails are just abut missing tags in the dataset card, merging !", "FYI, \r\n\r\nThe checksums have not been edited from the changes of .tgz to .zip files, and as a result a `ExpectedMoreDownloadedFiles` error occurs. Updating th...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4481", "html_url": "https://github.com/huggingface/datasets/pull/4481", "diff_url": "https://github.com/huggingface/datasets/pull/4481.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4481.patch", "merged_at": "2022-06-13T10:40...
4,481
true
Bigbench tensorflow GPU dependency
## Describe the bug Loading bigbech ```py from datasets import load_dataset dataset = load_dataset("bigbench","swedish_to_german_proverbs") ``` tries to use gpu and fails with OOM with the following error ``` Downloading and preparing dataset bigbench/swedish_to_german_proverbs (download: Unknown size, genera...
https://github.com/huggingface/datasets/issues/4480
[ "Thanks for reporting ! :) cc @andersjohanandreassen can you take a look at this ?\r\n\r\nAlso @cceyda feel free to open an issue at [BIG-Bench](https://github.com/google/BIG-bench) as well regarding the `AttributeError`", "I'm on vacation for the next week, so won't be able to do much debugging at the moment. So...
null
4,480
false
Include entity positions as feature in ReCoRD
https://huggingface.co/datasets/super_glue/viewer/record/validation TLDR: We need to record entity positions, which are included in the source data but excluded by the loading script, to enable efficient and effective training for ReCoRD. Currently, the loading script ignores the entity positions ("entity_start",...
https://github.com/huggingface/datasets/pull/4479
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks for the reply @lhoestq !\r\n\r\nI have sucessed on `datasets-cli test ./datasets/super_glue --name record --save_infos`,\r\nBut as you can see, the check ran into `FAILED tests/test_dataset_cards.py::test_changed_dataset_card[...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4479", "html_url": "https://github.com/huggingface/datasets/pull/4479", "diff_url": "https://github.com/huggingface/datasets/pull/4479.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4479.patch", "merged_at": "2022-08-19T13:23...
4,479
true
Dataset slow during model training
## Describe the bug While migrating towards πŸ€— Datasets, I encountered an odd performance degradation: training suddenly slows down dramatically. I train with an image dataset using Keras and execute a `to_tf_dataset` just before training. First, I have optimized my dataset following https://discuss.huggingface.co/...
https://github.com/huggingface/datasets/issues/4478
[ "Hi ! cc @Rocketknight1 maybe you know better ?\r\n\r\nI'm not too familiar with `tf.data.experimental.save`. Note that `datasets` uses memory mapping, so depending on your hardware and the disk you are using you can expect performance differences with a dataset loaded in RAM", "Hi @lehrig, I suspect what's happe...
null
4,478
false
Dataset Viewer issue for fgrezes/WIESP2022-NER
### Link _No response_ ### Description _No response_ ### Owner _No response_
https://github.com/huggingface/datasets/issues/4477
[ "https://huggingface.co/datasets/fgrezes/WIESP2022-NER\r\n\r\nThe error:\r\n\r\n```\r\nMessage: Couldn't find a dataset script at /src/services/worker/fgrezes/WIESP2022-NER/WIESP2022-NER.py or any data file in the same directory. Couldn't find 'fgrezes/WIESP2022-NER' on the Hugging Face Hub either: FileNotFou...
null
4,477
false
`to_pandas` doesn't take into account format.
**Is your feature request related to a problem? Please describe.** I have a large dataset that I need to convert part of to pandas to do some further analysis. Calling `to_pandas` directly on it is expensive. So I thought I could simply select the columns that I want and then call `to_pandas`. **Describe the solu...
https://github.com/huggingface/datasets/issues/4476
[ "Thanks for opening a discussion :)\r\n\r\nNote that you can use `.remove_columns(...)` to keep only the ones you're interested in before calling `.to_pandas()`", "Yes I can do that thank you!\r\n\r\nDo you think that conceptually my example should work? If not, I'm happy to close this issue. \r\n\r\nIf yes, I ca...
null
4,476
false
Improve error message for missing pacakges from inside dataset script
Improve the error message for missing packages from inside a dataset script: With this change, the error message for missing packages for `bigbench` looks as follows: ``` ImportError: To be able to use bigbench, you need to install the following dependencies: - 'bigbench' using 'pip install "bigbench @ ht...
https://github.com/huggingface/datasets/pull/4475
[ "_The documentation is not available anymore as the PR was closed or merged._", "I opened a PR before I noticed yours ^^' You can find it here: https://github.com/huggingface/datasets/pull/4484\r\n\r\nThe only comment I have regarding your message is that it possibly shows several `pip install` commands, whereas ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4475", "html_url": "https://github.com/huggingface/datasets/pull/4475", "diff_url": "https://github.com/huggingface/datasets/pull/4475.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4475.patch", "merged_at": null }
4,475
true
[Docs] How to use with PyTorch page
Currently the docs about PyTorch are scattered around different pages, and we were missing a place to explain more in depth how to use and optimize a dataset for PyTorch. This PR is related to #4457 which is the TF counterpart :) cc @Rocketknight1 we can try to align both documentations contents now I think cc @s...
https://github.com/huggingface/datasets/pull/4474
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4474", "html_url": "https://github.com/huggingface/datasets/pull/4474", "diff_url": "https://github.com/huggingface/datasets/pull/4474.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4474.patch", "merged_at": "2022-06-14T14:04...
4,474
true
Add SST-2 dataset
Add SST-2 dataset. Currently it is part of GLUE benchmark. This PR adds it as a standalone dataset. CC: @julien-c
https://github.com/huggingface/datasets/pull/4473
[ "_The documentation is not available anymore as the PR was closed or merged._", "on the hub this dataset is referenced as `sst-2` not `sst2` – is there a canonical orthography? If not, could we name it `sst-2`?", "@julien-c, we normally do not use hyphens for dataset names: whenever the original dataset name co...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4473", "html_url": "https://github.com/huggingface/datasets/pull/4473", "diff_url": "https://github.com/huggingface/datasets/pull/4473.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4473.patch", "merged_at": "2022-06-13T14:01...
4,473
true
Fix 401 error for unauthticated requests to non-existing repos
The hub now returns 401 instead of 404 for unauthenticated requests to non-existing repos. This PR add support for the 401 error and fixes the CI fails on `master`
https://github.com/huggingface/datasets/pull/4472
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4472", "html_url": "https://github.com/huggingface/datasets/pull/4472", "diff_url": "https://github.com/huggingface/datasets/pull/4472.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4472.patch", "merged_at": "2022-06-10T12:55...
4,472
true
CI error with repo lhoestq/_dummy
## Describe the bug CI is failing because of repo "lhoestq/_dummy". See: https://app.circleci.com/pipelines/github/huggingface/datasets/12461/workflows/1b040b45-9578-4ab9-8c44-c643c4eb8691/jobs/74269 ``` requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/api/datasets/lhoest...
https://github.com/huggingface/datasets/issues/4471
[ "fixed by https://github.com/huggingface/datasets/pull/4472" ]
null
4,471
false
Reorder returned validation/test splits in script template
null
https://github.com/huggingface/datasets/pull/4470
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4470", "html_url": "https://github.com/huggingface/datasets/pull/4470", "diff_url": "https://github.com/huggingface/datasets/pull/4470.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4470.patch", "merged_at": "2022-06-10T17:54...
4,470
true
Replace data URLs in wider_face dataset once hosted on the Hub
This PR replaces the URLs of data files in Google Drive with our Hub ones, once the data owners have approved to host their data on the Hub. They also informed us that their dataset is licensed under CC BY-NC-ND.
https://github.com/huggingface/datasets/pull/4469
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4469", "html_url": "https://github.com/huggingface/datasets/pull/4469", "diff_url": "https://github.com/huggingface/datasets/pull/4469.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4469.patch", "merged_at": "2022-06-10T16:32...
4,469
true
Generalize tutorials for audio and vision
This PR updates the tutorials to be more generalizable to all modalities. After reading the tutorials, a user should be able to load any type of dataset, know how to index into and slice a dataset, and do the most basic/common type of preprocessing (tokenization, resampling, applying transforms) depending on their data...
https://github.com/huggingface/datasets/pull/4468
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4468", "html_url": "https://github.com/huggingface/datasets/pull/4468", "diff_url": "https://github.com/huggingface/datasets/pull/4468.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4468.patch", "merged_at": "2022-06-14T16:12...
4,468
true
Transcript string 'null' converted to [None] by load_dataset()
## Issue I am training a luxembourgish speech-recognition model in Colab with a custom dataset, including a dictionary of luxembourgish words, for example the speaken numbers 0 to 9. When preparing the dataset with the script `ds_train1 = mydataset.map(prepare_dataset)` the following error was issued: ``` ...
https://github.com/huggingface/datasets/issues/4467
[ "Hi @mbarnig, thanks for reporting.\r\n\r\nPlease note that is an expected behavior by `pandas` (we use the `pandas` library to parse CSV files): https://pandas.pydata.org/docs/reference/api/pandas.read_csv.html\r\n```\r\nBy default the following values are interpreted as NaN: \r\nβ€˜β€™, β€˜#N/A’, β€˜#N/A N/A’, β€˜#NA’, β€˜-1...
null
4,467
false
Optimize contiguous shard and select
Currently `.shard()` and `.select()` always create an indices mapping. However if the requested data are contiguous, it's much more optimized to simply slice the Arrow table instead of building an indices mapping. In particular: - the shard/select operation will be much faster - reading speed will be much faster in t...
https://github.com/huggingface/datasets/pull/4466
[ "_The documentation is not available anymore as the PR was closed or merged._", "I thought of just mentioning the benefits I got. Here's the code that @lhoestq provided:\r\n\r\n```py\r\nimport os\r\nfrom datasets import load_dataset\r\nfrom tqdm.auto import tqdm\r\n\r\nds = load_dataset(\"squad\", split=\"train\"...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4466", "html_url": "https://github.com/huggingface/datasets/pull/4466", "diff_url": "https://github.com/huggingface/datasets/pull/4466.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4466.patch", "merged_at": "2022-06-14T15:54...
4,466
true
Fix bigbench config names
Fix https://github.com/huggingface/datasets/issues/4462 in the case of bigbench
https://github.com/huggingface/datasets/pull/4465
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4465", "html_url": "https://github.com/huggingface/datasets/pull/4465", "diff_url": "https://github.com/huggingface/datasets/pull/4465.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4465.patch", "merged_at": "2022-06-09T14:29...
4,465
true
Extend support for streaming datasets that use xml.dom.minidom.parse
This PR extends the support in streaming mode for datasets that use `xml.dom.minidom.parse`, by patching that function. This PR adds support for streaming datasets like "Yaxin/SemEval2015". Fix #4453.
https://github.com/huggingface/datasets/pull/4464
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4464", "html_url": "https://github.com/huggingface/datasets/pull/4464", "diff_url": "https://github.com/huggingface/datasets/pull/4464.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4464.patch", "merged_at": "2022-06-09T08:34...
4,464
true