title stringlengths 1 290 | body stringlengths 0 228k ⌀ | html_url stringlengths 46 51 | comments list | pull_request dict | number int64 1 5.59k | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|
Docs for creating an audio dataset | This PR is a first draft of how to create audio datasets (`AudioFolder` and loading script). Feel free to let me know if there are any specificities I'm missing for this. 🙂 | https://github.com/huggingface/datasets/pull/4872 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Awesome thanks ! I think we can also encourage TAR archives as for image dataset scripts (feel free to copy paste some parts from there lol)",
"Thanks for all the great feedback @polinaeterna and @lhoestq! 🥰\r\n\r\nI added all the... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4872",
"html_url": "https://github.com/huggingface/datasets/pull/4872",
"diff_url": "https://github.com/huggingface/datasets/pull/4872.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4872.patch",
"merged_at": "2022-09-21T10:27... | 4,872 | true |
Fix: wmt datasets - fix CWMT zh subsets | Fix https://github.com/huggingface/datasets/issues/4575
TODO: run `datasets-cli test`:
- [x] wmt17
- [x] wmt18
- [x] wmt19 | https://github.com/huggingface/datasets/pull/4871 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4871). All of your documentation changes will be reflected on that endpoint."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4871",
"html_url": "https://github.com/huggingface/datasets/pull/4871",
"diff_url": "https://github.com/huggingface/datasets/pull/4871.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4871.patch",
"merged_at": "2022-08-23T10:00... | 4,871 | true |
audio folder check CI | null | https://github.com/huggingface/datasets/pull/4870 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4870",
"html_url": "https://github.com/huggingface/datasets/pull/4870",
"diff_url": "https://github.com/huggingface/datasets/pull/4870.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4870.patch",
"merged_at": null
} | 4,870 | true |
Fix typos in documentation | null | https://github.com/huggingface/datasets/pull/4869 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4869",
"html_url": "https://github.com/huggingface/datasets/pull/4869",
"diff_url": "https://github.com/huggingface/datasets/pull/4869.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4869.patch",
"merged_at": "2022-08-22T09:09... | 4,869 | true |
adding mafand to datasets | I'm addding the MAFAND dataset by Masakhane based on the paper/repository below:
Paper: https://aclanthology.org/2022.naacl-main.223/
Code: https://github.com/masakhane-io/lafand-mt
Please, help merge this
Everything works except for creating dummy data file | https://github.com/huggingface/datasets/pull/4868 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Hi @dadelani, thanks for your awesome contribution!!! :heart: \r\n\r\nHowever, now we are using the Hub to add new datasets, instead of this GitHub repo. \r\n\r\nYou could share this dataset under your Hub organization namespace: [Ma... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4868",
"html_url": "https://github.com/huggingface/datasets/pull/4868",
"diff_url": "https://github.com/huggingface/datasets/pull/4868.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4868.patch",
"merged_at": null
} | 4,868 | true |
Complete tags of superglue dataset card | Related to #4479 . | https://github.com/huggingface/datasets/pull/4867 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4867",
"html_url": "https://github.com/huggingface/datasets/pull/4867",
"diff_url": "https://github.com/huggingface/datasets/pull/4867.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4867.patch",
"merged_at": "2022-08-22T08:58... | 4,867 | true |
amend docstring for dunder | display dunder method in docsting with underlines an not bold markdown. | https://github.com/huggingface/datasets/pull/4866 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4866). All of your documentation changes will be reflected on that endpoint."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4866",
"html_url": "https://github.com/huggingface/datasets/pull/4866",
"diff_url": "https://github.com/huggingface/datasets/pull/4866.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4866.patch",
"merged_at": null
} | 4,866 | true |
Dataset Viewer issue for MoritzLaurer/multilingual_nli | ### Link
_No response_
### Description
I've just uploaded a new dataset to the hub and the viewer does not work for some reason, see here: https://huggingface.co/datasets/MoritzLaurer/multilingual_nli
It displays the error:
```
Status code: 400
Exception: Status400Error
Message: The dataset... | https://github.com/huggingface/datasets/issues/4865 | [
"Thanks for reporting @MoritzLaurer.\r\n\r\nCurrently, the dataset preview is working properly: https://huggingface.co/datasets/MoritzLaurer/multilingual_nli\r\n\r\nPlease note that when a dataset is modified, it might take some time until the preview is completely updated.\r\n\r\n@severo might it be worth adding ... | null | 4,865 | false |
Allow pathlib PoxisPath in Dataset.read_json | **Is your feature request related to a problem? Please describe.**
```
from pathlib import Path
from datasets import Dataset
ds = Dataset.read_json(Path('data.json'))
```
causes an error
```
AttributeError: 'PosixPath' object has no attribute 'decode'
```
**Describe the solution you'd like**
It should be... | https://github.com/huggingface/datasets/issues/4864 | [
"This same error will occur using `ds = datasets.load_dataset('json', data_files=['test.jsonl'])`",
"@cccntu I want to make a quick fix for this, but I am struggling to find where the json dataset builder is. Do you know?",
"@vvvm23 I think you mean think:\r\n```python\r\nds = datasets.load_dataset('json', data... | null | 4,864 | false |
TFDS wiki_dialog dataset to Huggingface dataset | ## Adding a Dataset
- **Name:** *Wiki_dialog*
- **Description: https://github.com/google-research/dialog-inpainting#:~:text=JSON%20object%2C%20for-,example,-%3A
- **Paper: https://arxiv.org/abs/2205.09073
- **Data: https://github.com/google-research/dialog-inpainting
- **Motivation:** *Research and Development on ... | https://github.com/huggingface/datasets/issues/4863 | [
"@albertvillanova any help ? The linked dataset is in beam format which is similar to wikipedia dataset in huggingface that you scripted..",
"Nvm, I was able to port it to huggingface datasets, will upload to the hub soon",
"https://huggingface.co/datasets/djaym7/wiki_dialog",
"Thanks for the addition, @djay... | null | 4,863 | false |
Got "AttributeError: 'xPath' object has no attribute 'read'" when loading an excel dataset with my own code | ## Describe the bug
A clear and concise description of what the bug is.
## Steps to reproduce the bug
```python
# Sample code to reproduce the bug
# The dataset function is as follows:
from pathlib import Path
from typing import Dict, List, Tuple
import datasets
import pandas as pd
_CITATION = """\
"""... | https://github.com/huggingface/datasets/issues/4862 | [
"What's more, the downloaded data is actually a folder instead of an excel file.",
"Hi hi, instead of using `download_and_extract` function, I only use `download` function: `base_dir = Path(dl_manager.download(urls))`. It turns out that the code works for `datasets==2.2.2`, however, it doesn't work with `datasets... | null | 4,862 | false |
Using disk for memory with the method `from_dict` | **Is your feature request related to a problem? Please describe.**
I start with an empty dataset. In a loop, at each iteration, I create a new dataset with the method `from_dict` (based on some data I load) and I concatenate this new dataset with the one at the previous iteration. After some iterations, I have an OOM ... | https://github.com/huggingface/datasets/issues/4861 | [
"This issue was also causing an OOM in @nateraw 's workflow and shows again that behavior is confusing - we should definitely switch to using the disk IMO"
] | null | 4,861 | false |
Add collection3 dataset | null | https://github.com/huggingface/datasets/pull/4860 | [
"Hi @pefimov. Thanks for you awesome work on this dataset contribution.\r\n\r\nHowever, now we are using the Hub to add new datasets, instead of this GitHub repo. \r\n\r\nYou could share this dataset under the appropriate Hub organization namespace. This way the dataset will be accessible using:\r\n```python\r\nds ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4860",
"html_url": "https://github.com/huggingface/datasets/pull/4860",
"diff_url": "https://github.com/huggingface/datasets/pull/4860.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4860.patch",
"merged_at": null
} | 4,860 | true |
can't install using conda on Windows 10 | ## Describe the bug
I wanted to install using conda or Anaconda navigator. That didn't work, so I had to install using pip.
## Steps to reproduce the bug
conda install -c huggingface -c conda-forge datasets
## Expected results
Should have indicated successful installation.
## Actual results
Solving environ... | https://github.com/huggingface/datasets/issues/4859 | [] | null | 4,859 | false |
map() function removes columns when input_columns is not None | ## Describe the bug
The map function, removes features from the dataset that are not present in the _input_columns_ list of columns, despite the columns being removed not mentioned in the _remove_columns_ argument.
## Steps to reproduce the bug
```python
from datasets import Dataset
ds = Dataset.from_dict({"a" : [... | https://github.com/huggingface/datasets/issues/4858 | [
"Hi! Thanks for reporting! This looks like a bug. I've just opened a PR with the fix.",
"Awesome! Thank you. I'll close the issue once the PR gets merged. :-)",
"I guess we should reopen after the revert by:\r\n- #5006"
] | null | 4,858 | false |
No preprocessed wikipedia is working on huggingface/datasets | ## Describe the bug
20220301 wikipedia dump has been deprecated, so now there is no working wikipedia dump on huggingface
https://huggingface.co/datasets/wikipedia
https://dumps.wikimedia.org/enwiki/
| https://github.com/huggingface/datasets/issues/4857 | [
"Thanks for reporting @aninrusimha.\r\n\r\nPlease, note that the preprocessed datasets are still available, as described in the dataset card, e.g.: https://huggingface.co/datasets/wikipedia\r\n```python\r\nds = load_dataset(\"wikipedia\", \"20220301.en\")\r\n``` ",
"This is working now, but I was getting an error... | null | 4,857 | false |
file missing when load_dataset with openwebtext on windows | ## Describe the bug
0015896-b1054262f7da52a0518521e29c8e352c.txt is missing when I run run_mlm.py with openwebtext. I check the cache_path and can not find 0015896-b1054262f7da52a0518521e29c8e352c.txt. but I can find this file in the 17ecf461bfccd469a1fbc264ccb03731f8606eea7b3e2e8b86e13d18040bf5b3/urlsf_subset00-16_da... | https://github.com/huggingface/datasets/issues/4856 | [
"I have tried to extract ```0015896-b1054262f7da52a0518521e29c8e352c.txt``` from ```17ecf461bfccd469a1fbc264ccb03731f8606eea7b3e2e8b86e13d18040bf5b3/urlsf_subset00-16_data.xz``` with 7-zip\r\nand put the file into cache_path ```F://huggingface/datasets/downloads/extracted/0901d27f43b7e9ac0577da0d0061c8c632ba0b70ecd... | null | 4,856 | false |
Dataset Viewer issue for super_glue | ### Link
https://huggingface.co/datasets/super_glue
### Description
can't view super_glue dataset on the web page
### Owner
_No response_ | https://github.com/huggingface/datasets/issues/4855 | [
"Thanks for reporting @wzsxxa.\r\n\r\nHowever the \"super_glue\" dataset is rendered properly by the Dataset preview: https://huggingface.co/datasets/super_glue"
] | null | 4,855 | false |
Fix bug and checksums in exams dataset | Fix #4852. | https://github.com/huggingface/datasets/pull/4853 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4853",
"html_url": "https://github.com/huggingface/datasets/pull/4853",
"diff_url": "https://github.com/huggingface/datasets/pull/4853.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4853.patch",
"merged_at": "2022-08-16T06:29... | 4,853 | true |
Bug in multilingual_with_para config of exams dataset and checksums error | ## Describe the bug
There is a bug for "multilingual_with_para" config in exams dataset:
```python
ds = load_dataset("./datasets/exams", split="train")
```
raises:
```
KeyError: 'choices'
```
Moreover, there is a NonMatchingChecksumError:
```
NonMatchingChecksumError: Checksums didn't match for dataset so... | https://github.com/huggingface/datasets/issues/4852 | [
"Hi @albertvillanova. Unfortunately I still get this error. Is this because the merge has yet to be released? Is there a way to track the release?",
"Hi @thesofakillers, yes you are right: the fix will be available after next release (it was planned for today; Monday at the latest).\r\n\r\nIn the meantime, you ca... | null | 4,852 | false |
Fix license tag and Source Data section in billsum dataset card | Fixed the data source and license fields | https://github.com/huggingface/datasets/pull/4851 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"thanks @albertvillanova done thank you!"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4851",
"html_url": "https://github.com/huggingface/datasets/pull/4851",
"diff_url": "https://github.com/huggingface/datasets/pull/4851.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4851.patch",
"merged_at": "2022-08-22T13:40... | 4,851 | true |
Fix test of _get_extraction_protocol for TAR files | While working in another PR, I discovered an xpass test (a test that is supposed to xfail but nevertheless passes) when testing `_get_extraction_protocol`: https://github.com/huggingface/datasets/runs/7818845285?check_suite_focus=true
```
XPASS tests/test_streaming_download_manager.py::test_streaming_dl_manager_get_e... | https://github.com/huggingface/datasets/pull/4850 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4850",
"html_url": "https://github.com/huggingface/datasets/pull/4850",
"diff_url": "https://github.com/huggingface/datasets/pull/4850.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4850.patch",
"merged_at": "2022-08-15T09:28... | 4,850 | true |
1.18.x | null | https://github.com/huggingface/datasets/pull/4849 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4849",
"html_url": "https://github.com/huggingface/datasets/pull/4849",
"diff_url": "https://github.com/huggingface/datasets/pull/4849.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4849.patch",
"merged_at": null
} | 4,849 | true |
a | null | https://github.com/huggingface/datasets/pull/4848 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4848",
"html_url": "https://github.com/huggingface/datasets/pull/4848",
"diff_url": "https://github.com/huggingface/datasets/pull/4848.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4848.patch",
"merged_at": null
} | 4,848 | true |
Test win ci | aa | https://github.com/huggingface/datasets/pull/4847 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4847",
"html_url": "https://github.com/huggingface/datasets/pull/4847",
"diff_url": "https://github.com/huggingface/datasets/pull/4847.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4847.patch",
"merged_at": null
} | 4,847 | true |
Update documentation card of miam dataset | Hi !
Paper has been published at EMNLP. | https://github.com/huggingface/datasets/pull/4846 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Ahahah :D not sur how i broke something by updating the README :D ",
"Thanks for the fix @PierreColombo. \r\n\r\nOnce a README is modified, our CI runs tests on it, requiring additional quality fixes, so that all READMEs are progre... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4846",
"html_url": "https://github.com/huggingface/datasets/pull/4846",
"diff_url": "https://github.com/huggingface/datasets/pull/4846.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4846.patch",
"merged_at": "2022-08-14T10:26... | 4,846 | true |
Mark CI tests as xfail if Hub HTTP error | In order to make testing more robust (and avoid merges to master with red tests), we could mark tests as xfailed (instead of failed) when the Hub raises some temporary HTTP errors.
This PR:
- marks tests as xfailed only if the Hub raises a 500 error for:
- test_upstream_hub
- makes pytest report the xfailed/xpa... | https://github.com/huggingface/datasets/pull/4845 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4845",
"html_url": "https://github.com/huggingface/datasets/pull/4845",
"diff_url": "https://github.com/huggingface/datasets/pull/4845.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4845.patch",
"merged_at": "2022-08-23T04:42... | 4,845 | true |
Add 'val' to VALIDATION_KEYWORDS. | This PR fixes #4839 by adding the word `"val"` to the `VALIDATION_KEYWORDS` so that the `load_dataset()` method with `imagefolder` (and probably, some other directives as well) reads folders named `"val"` as well.
I think the supported keywords have to be mentioned in the documentation as well, but I couldn't think ... | https://github.com/huggingface/datasets/pull/4844 | [
"@mariosasko not sure about how the reviewing process works. Maybe you can have a look because we discussed this elsewhere?",
"Hi, thanks! \r\n\r\nLet's add one pattern with `val` to this test before merging: \r\nhttps://github.com/huggingface/datasets/blob/b88a656cf94c4ad972154371c83c1af759fde522/tests/test_data... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4844",
"html_url": "https://github.com/huggingface/datasets/pull/4844",
"diff_url": "https://github.com/huggingface/datasets/pull/4844.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4844.patch",
"merged_at": "2022-08-30T10:14... | 4,844 | true |
Fix typo in streaming docs | null | https://github.com/huggingface/datasets/pull/4843 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4843",
"html_url": "https://github.com/huggingface/datasets/pull/4843",
"diff_url": "https://github.com/huggingface/datasets/pull/4843.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4843.patch",
"merged_at": "2022-08-14T11:02... | 4,843 | true |
Update stackexchange license | The correct license of the stackexchange subset of the Pile is `cc-by-sa-4.0`, as can for example be seen here: https://stackoverflow.com/help/licensing | https://github.com/huggingface/datasets/pull/4842 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4842",
"html_url": "https://github.com/huggingface/datasets/pull/4842",
"diff_url": "https://github.com/huggingface/datasets/pull/4842.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4842.patch",
"merged_at": "2022-08-14T10:28... | 4,842 | true |
Update ted_talks_iwslt license to include ND | Excerpt from the paper's abstract: "Aside from its cultural and social relevance, this content, which is published under the Creative Commons BY-NC-ND license, also represents a precious language resource for the machine translation research community" | https://github.com/huggingface/datasets/pull/4841 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4841",
"html_url": "https://github.com/huggingface/datasets/pull/4841",
"diff_url": "https://github.com/huggingface/datasets/pull/4841.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4841.patch",
"merged_at": "2022-08-14T11:00... | 4,841 | true |
Dataset Viewer issue for darragh/demo_data_raw3 | ### Link
https://huggingface.co/datasets/darragh/demo_data_raw3
### Description
```
Exception: ValueError
Message: Arrow type extension<arrow.py_extension_type<pyarrow.lib.UnknownExtensionType>> does not have a datasets dtype equivalent.
```
reported by @NielsRogge
### Owner
No | https://github.com/huggingface/datasets/issues/4840 | [
"do you have an idea of why it can occur @huggingface/datasets? The dataset consists of a single parquet file.",
"Thanks for reporting @severo.\r\n\r\nI'm not able to reproduce that error. I get instead:\r\n```\r\nFileNotFoundError: [Errno 2] No such file or directory: 'orix/data/ChiSig/唐合乐-9-3.jpg'\r\n```\r\n\r\... | null | 4,840 | false |
ImageFolder dataset builder does not read the validation data set if it is named as "val" | **Is your feature request related to a problem? Please describe.**
Currently, the `'imagefolder'` data set builder in [`load_dataset()`](https://github.com/huggingface/datasets/blob/2.4.0/src/datasets/load.py#L1541] ) only [supports](https://github.com/huggingface/datasets/blob/6c609a322da994de149b2c938f19439bca9940... | https://github.com/huggingface/datasets/issues/4839 | [
"#take"
] | null | 4,839 | false |
Fix documentation card of adv_glue dataset | Fix documentation card of adv_glue dataset. | https://github.com/huggingface/datasets/pull/4838 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The failing test has nothing to do with this PR:\r\n```\r\nFAILED tests/test_upstream_hub.py::TestPushToHub::test_push_dataset_dict_to_hub_multiple_files\r\n```"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4838",
"html_url": "https://github.com/huggingface/datasets/pull/4838",
"diff_url": "https://github.com/huggingface/datasets/pull/4838.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4838.patch",
"merged_at": "2022-08-15T10:02... | 4,838 | true |
Add support for CSV metadata files to ImageFolder | Fix #4814 | https://github.com/huggingface/datasets/pull/4837 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Cool thanks ! Maybe let's include this change after the refactoring from FolderBasedBuilder in #3963 to avoid dealing with too many unpleasant conflicts ?",
"@lhoestq I resolved the conflicts (AudioFolder also supports CSV metadata... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4837",
"html_url": "https://github.com/huggingface/datasets/pull/4837",
"diff_url": "https://github.com/huggingface/datasets/pull/4837.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4837.patch",
"merged_at": "2022-08-31T11:59... | 4,837 | true |
Is it possible to pass multiple links to a split in load script? | **Is your feature request related to a problem? Please describe.**
I wanted to use a python loading script in hugging face datasets that use different sources of text (it's somehow a compilation of multiple datasets + my own dataset) based on how `load_dataset` [works](https://huggingface.co/docs/datasets/loading) I a... | https://github.com/huggingface/datasets/issues/4836 | [] | null | 4,836 | false |
Fix documentation card of ethos dataset | Fix documentation card of ethos dataset. | https://github.com/huggingface/datasets/pull/4835 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4835",
"html_url": "https://github.com/huggingface/datasets/pull/4835",
"diff_url": "https://github.com/huggingface/datasets/pull/4835.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4835.patch",
"merged_at": "2022-08-12T12:59... | 4,835 | true |
Fix documentation card of recipe_nlg dataset | Fix documentation card of recipe_nlg dataset | https://github.com/huggingface/datasets/pull/4834 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4834",
"html_url": "https://github.com/huggingface/datasets/pull/4834",
"diff_url": "https://github.com/huggingface/datasets/pull/4834.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4834.patch",
"merged_at": "2022-08-12T11:13... | 4,834 | true |
Fix missing tags in dataset cards | Fix missing tags in dataset cards:
- boolq
- break_data
- definite_pronoun_resolution
- emo
- kor_nli
- pg19
- quartz
- sciq
- squad_es
- wmt14
- wmt15
- wmt16
- wmt17
- wmt18
- wmt19
- wmt_t2t
This PR partially fixes the missing tags in dataset cards. Subsequent PRs will follow to complete this task... | https://github.com/huggingface/datasets/pull/4833 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4833",
"html_url": "https://github.com/huggingface/datasets/pull/4833",
"diff_url": "https://github.com/huggingface/datasets/pull/4833.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4833.patch",
"merged_at": "2022-08-12T09:45... | 4,833 | true |
Fix tags in dataset cards | Fix wrong tags in dataset cards. | https://github.com/huggingface/datasets/pull/4832 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The non-passing tests are caused by other missing information in the dataset cards."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4832",
"html_url": "https://github.com/huggingface/datasets/pull/4832",
"diff_url": "https://github.com/huggingface/datasets/pull/4832.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4832.patch",
"merged_at": "2022-08-12T04:27... | 4,832 | true |
Add oversampling strategies to interleave datasets | Hello everyone,
Here is a proposal to improve `interleave_datasets` function.
Following Issue #3064, and @lhoestq [comment](https://github.com/huggingface/datasets/issues/3064#issuecomment-1022333385), I propose here a code that performs oversampling when interleaving a `Dataset` list.
I have myself encountered t... | https://github.com/huggingface/datasets/pull/4831 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4831). All of your documentation changes will be reflected on that endpoint.",
"Hi @lhoestq, \r\nThanks for your review! I've added the requested mention in the documentation and corrected the Error type in `interleave_datasets... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4831",
"html_url": "https://github.com/huggingface/datasets/pull/4831",
"diff_url": "https://github.com/huggingface/datasets/pull/4831.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4831.patch",
"merged_at": "2022-08-24T16:46... | 4,831 | true |
Fix task tags in dataset cards | null | https://github.com/huggingface/datasets/pull/4830 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The non-passing tests are caused by other missing information in the dataset cards."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4830",
"html_url": "https://github.com/huggingface/datasets/pull/4830",
"diff_url": "https://github.com/huggingface/datasets/pull/4830.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4830.patch",
"merged_at": "2022-08-11T16:23... | 4,830 | true |
Misalignment between card tag validation and docs | ## Describe the bug
As pointed out in other issue: https://github.com/huggingface/datasets/pull/4827#discussion_r943536284
the validation of the dataset card tags is not aligned with its documentation: e.g.
- implementation: `license: List[str]`
- docs: `license: Union[str, List[str]]`
They should be aligned.
... | https://github.com/huggingface/datasets/issues/4829 | [
"(Note that the doc is aligned with the hub validation rules, and the \"ground truth\" is the hub validation rules given that they apply to all datasets, not just the canonical ones)"
] | null | 4,829 | false |
Support PIL Image objects in `add_item`/`add_column` | Fix #4796
PS: We should also improve the type inference in `OptimizedTypeSequence` to make it possible to also infer the complex types (only `Image` currently) in nested arrays (e.g. `[[pil_image], [pil_image, pil_image]]` or `[{"img": pil_image}`]), but I plan to address this in a separate PR. | https://github.com/huggingface/datasets/pull/4828 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4828). All of your documentation changes will be reflected on that endpoint.",
"Hey @mariosasko could we please merge this? I'm still getting the original error at #4796 ."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4828",
"html_url": "https://github.com/huggingface/datasets/pull/4828",
"diff_url": "https://github.com/huggingface/datasets/pull/4828.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4828.patch",
"merged_at": null
} | 4,828 | true |
Add license metadata to pg19 | As reported over email by Roy Rijkers | https://github.com/huggingface/datasets/pull/4827 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4827",
"html_url": "https://github.com/huggingface/datasets/pull/4827",
"diff_url": "https://github.com/huggingface/datasets/pull/4827.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4827.patch",
"merged_at": "2022-08-11T14:46... | 4,827 | true |
Fix language tags in dataset cards | Fix language tags in all dataset cards, so that they are validated (aligned with our `languages.json` resource). | https://github.com/huggingface/datasets/pull/4826 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The non-passing tests are caused by other missing information in the dataset cards."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4826",
"html_url": "https://github.com/huggingface/datasets/pull/4826",
"diff_url": "https://github.com/huggingface/datasets/pull/4826.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4826.patch",
"merged_at": "2022-08-11T14:03... | 4,826 | true |
[Windows] Fix Access Denied when using os.rename() | In this PR, we are including an additional step when `os.rename()` raises a PermissionError.
Basically, we will use `shutil.move()` on the temp files.
Fix #2937 | https://github.com/huggingface/datasets/pull/4825 | [
"Cool thank you ! Maybe we can just replace `os.rename` by `shutil.move` instead ?",
"> Cool thank you ! Maybe we can just replace `os.rename` by `shutil.move` instead ?\r\n\r\nYes, I think that could be a better solution, but I didn't test it in Linux (e.g. Ubuntu) to guarantee that `os.rename()` could be comple... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4825",
"html_url": "https://github.com/huggingface/datasets/pull/4825",
"diff_url": "https://github.com/huggingface/datasets/pull/4825.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4825.patch",
"merged_at": "2022-08-24T13:09... | 4,825 | true |
Fix titles in dataset cards | Fix all the titles in the dataset cards, so that they conform to the required format. | https://github.com/huggingface/datasets/pull/4824 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The non-passing tests are caused by other missing information in the dataset cards."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4824",
"html_url": "https://github.com/huggingface/datasets/pull/4824",
"diff_url": "https://github.com/huggingface/datasets/pull/4824.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4824.patch",
"merged_at": "2022-08-11T12:56... | 4,824 | true |
Update data URL in mkqa dataset | Update data URL in mkqa dataset.
Fix #4817. | https://github.com/huggingface/datasets/pull/4823 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4823",
"html_url": "https://github.com/huggingface/datasets/pull/4823",
"diff_url": "https://github.com/huggingface/datasets/pull/4823.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4823.patch",
"merged_at": "2022-08-11T09:37... | 4,823 | true |
Moving dataset between namespaces breaks dataset viewer | ## Describe the bug
I moved a dataset from my own namespace to an org and that broke the dataset viewer. To fix it I had to manually edit the `dataset_info.json` file and change the first key in the json from `username--datasetname` to `orgname--datasetname`
## Steps to reproduce the bug
What I did was:
1- Uploa... | https://github.com/huggingface/datasets/issues/4822 | [
"Let's keep open for now. We should try to reproduce"
] | null | 4,822 | false |
Fix train_test_split docs | I saw that `stratify` is added to the `train_test_split` method as per #4322, hence the docs can be updated. | https://github.com/huggingface/datasets/pull/4821 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4821",
"html_url": "https://github.com/huggingface/datasets/pull/4821",
"diff_url": "https://github.com/huggingface/datasets/pull/4821.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4821.patch",
"merged_at": "2022-08-11T09:45... | 4,821 | true |
Terminating: fork() called from a process already using GNU OpenMP, this is unsafe. | Hi, when i try to run prepare_dataset function in [fine tuning ASR tutorial 4](https://colab.research.google.com/github/patrickvonplaten/notebooks/blob/master/Fine_tuning_Wav2Vec2_for_English_ASR.ipynb) , i got this error.
I got this error
Terminating: fork() called from a process already using GNU OpenMP, this is un... | https://github.com/huggingface/datasets/issues/4820 | [
"Fixed by installing either resampy<3 or resampy>=4"
] | null | 4,820 | false |
Add missing language tags to resources | Add missing language tags to resources, required by existing datasets on GitHub. | https://github.com/huggingface/datasets/pull/4819 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4819",
"html_url": "https://github.com/huggingface/datasets/pull/4819",
"diff_url": "https://github.com/huggingface/datasets/pull/4819.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4819.patch",
"merged_at": "2022-08-10T19:32... | 4,819 | true |
Add add cc-by-sa-2.5 license tag | - [ ] add it to moon-landing
- [ ] add it to hub-docs | https://github.com/huggingface/datasets/pull/4818 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4818). All of your documentation changes will be reflected on that endpoint.",
"I think we can close this PR because the `standard_licenses.tsv` file was removed from this repo and we no longer perform any dataset card validati... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4818",
"html_url": "https://github.com/huggingface/datasets/pull/4818",
"diff_url": "https://github.com/huggingface/datasets/pull/4818.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4818.patch",
"merged_at": null
} | 4,818 | true |
Outdated Link for mkqa Dataset | ## Describe the bug
The URL used to download the mkqa dataset is outdated. It seems the URL to download the dataset is currently https://github.com/apple/ml-mkqa/blob/main/dataset/mkqa.jsonl.gz instead of https://github.com/apple/ml-mkqa/raw/master/dataset/mkqa.jsonl.gz (master branch has been renamed to main).
## ... | https://github.com/huggingface/datasets/issues/4817 | [
"Thanks for reporting @liaeh, we are investigating this. "
] | null | 4,817 | false |
Update version of opus_paracrawl dataset | This PR updates OPUS ParaCrawl from 7.1 to 9 version.
Fix #4815. | https://github.com/huggingface/datasets/pull/4816 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4816",
"html_url": "https://github.com/huggingface/datasets/pull/4816",
"diff_url": "https://github.com/huggingface/datasets/pull/4816.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4816.patch",
"merged_at": "2022-08-12T14:17... | 4,816 | true |
Outdated loading script for OPUS ParaCrawl dataset | ## Describe the bug
Our loading script for OPUS ParaCrawl loads its 7.1 version. Current existing version is 9.
| https://github.com/huggingface/datasets/issues/4815 | [] | null | 4,815 | false |
Support CSV as metadata file format in AudioFolder/ImageFolder | Requested here: https://discuss.huggingface.co/t/how-to-structure-an-image-dataset-repo-using-the-image-folder-approach/21004. CSV is also used in AutoTrain for specifying metadata in image datasets. | https://github.com/huggingface/datasets/issues/4814 | [] | null | 4,814 | false |
Fix loading example in opus dataset cards | This PR:
- fixes the examples to load the datasets, with the corrected dataset name, in their dataset cards for:
- opus_dgt
- opus_paracrawl
- opus_wikipedia
- fixes their dataset cards with the missing required information: title, data instances/fields/splits
- enumerates the supported languages
- adds a ... | https://github.com/huggingface/datasets/pull/4813 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4813",
"html_url": "https://github.com/huggingface/datasets/pull/4813",
"diff_url": "https://github.com/huggingface/datasets/pull/4813.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4813.patch",
"merged_at": "2022-08-09T17:38... | 4,813 | true |
Fix bug in function validate_type for Python >= 3.9 | Fix `validate_type` function, so that it uses `get_origin` instead. This makes the function forward compatible.
This fixes #4811 because:
```python
In [4]: typing.Optional[str]
Out[4]: typing.Optional[str]
In [5]: get_origin(typing.Optional[str])
Out[5]: typing.Union
```
Fix #4811. | https://github.com/huggingface/datasets/pull/4812 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4812",
"html_url": "https://github.com/huggingface/datasets/pull/4812",
"diff_url": "https://github.com/huggingface/datasets/pull/4812.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4812.patch",
"merged_at": "2022-08-12T13:27... | 4,812 | true |
Bug in function validate_type for Python >= 3.9 | ## Describe the bug
The function `validate_type` assumes that the type `typing.Optional[str]` is automatically transformed to `typing.Union[str, NoneType]`.
```python
In [4]: typing.Optional[str]
Out[4]: typing.Union[str, NoneType]
```
However, this is not the case for Python 3.9:
```python
In [3]: typing.Opt... | https://github.com/huggingface/datasets/issues/4811 | [] | null | 4,811 | false |
Add description to hellaswag dataset | null | https://github.com/huggingface/datasets/pull/4810 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Are the `metadata JSON file` not on their way to deprecation? 😆😇\r\n\r\nIMO, more generally than this particular PR, the contribution process should be simplified now that many validation checks happen on the hub side.\r\n\r\nKeepi... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4810",
"html_url": "https://github.com/huggingface/datasets/pull/4810",
"diff_url": "https://github.com/huggingface/datasets/pull/4810.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4810.patch",
"merged_at": "2022-09-23T11:33... | 4,810 | true |
Complete the mlqa dataset card | I fixed the issue #4808
Details of PR:
- Added languages included in the dataset.
- Added task id and task category.
- Updated the citation information.
Fix #4808. | https://github.com/huggingface/datasets/pull/4809 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"> Thanks for your contribution, @eldhoittangeorge.\r\n> \r\n> The CI error message: https://github.com/huggingface/datasets/runs/7743526624?check_suite_focus=true\r\n> \r\n> ```\r\n> E ValueError: The following issues have ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4809",
"html_url": "https://github.com/huggingface/datasets/pull/4809",
"diff_url": "https://github.com/huggingface/datasets/pull/4809.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4809.patch",
"merged_at": "2022-08-09T13:26... | 4,809 | true |
Add more information to the dataset card of mlqa dataset | null | https://github.com/huggingface/datasets/issues/4808 | [
"#self-assign",
"Fixed by:\r\n- #4809"
] | null | 4,808 | false |
document fix in opus_gnome dataset | I fixed a issue #4805.
I changed `"gnome"` to `"opus_gnome"` in[ README.md](https://github.com/huggingface/datasets/tree/main/datasets/opus_gnome#dataset-summary). | https://github.com/huggingface/datasets/pull/4807 | [
"Duplicate:\r\n- #4806 "
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4807",
"html_url": "https://github.com/huggingface/datasets/pull/4807",
"diff_url": "https://github.com/huggingface/datasets/pull/4807.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4807.patch",
"merged_at": null
} | 4,807 | true |
Fix opus_gnome dataset card | I fixed a issue #4805.
I changed `"gnome"` to `"opus_gnome"` in[ README.md](https://github.com/huggingface/datasets/tree/main/datasets/opus_gnome#dataset-summary).
Fix #4805 | https://github.com/huggingface/datasets/pull/4806 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@gojiteji why have you closed this PR and created an identical one?\r\n- #4807 ",
"@albertvillanova \r\nI forgot to follow \"How to create a Pull\" in CONTRIBUTING.md in this branch.",
"Both are identical. And you can push additi... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4806",
"html_url": "https://github.com/huggingface/datasets/pull/4806",
"diff_url": "https://github.com/huggingface/datasets/pull/4806.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4806.patch",
"merged_at": "2022-08-09T11:52... | 4,806 | true |
Wrong example in opus_gnome dataset card | ## Describe the bug
I found that [the example on opus_gone dataset ](https://github.com/huggingface/datasets/tree/main/datasets/opus_gnome#dataset-summary) doesn't work.
## Steps to reproduce the bug
```python
load_dataset("gnome", lang1="it", lang2="pl")
```
`"gnome"` should be `"opus_gnome"`
## Expected r... | https://github.com/huggingface/datasets/issues/4805 | [] | null | 4,805 | false |
streaming dataset with concatenating splits raises an error | ## Describe the bug
streaming dataset with concatenating splits raises an error
## Steps to reproduce the bug
```python
from datasets import load_dataset
# no error
repo = "nateraw/ade20k-tiny"
dataset = load_dataset(repo, split="train+validation")
```
```python
from datasets import load_dataset
# er... | https://github.com/huggingface/datasets/issues/4804 | [
"Hi! Only the name of a particular split (\"train\", \"test\", ...) is supported as a split pattern if `streaming=True`. We plan to address this limitation soon."
] | null | 4,804 | false |
Support `pipeline` argument in inspect.py functions | **Is your feature request related to a problem? Please describe.**
The `wikipedia` dataset requires a `pipeline` argument to build the list of splits:
https://huggingface.co/datasets/wikipedia/blob/main/wikipedia.py#L937
But this is currently not supported in `get_dataset_config_info`:
https://github.com/hu... | https://github.com/huggingface/datasets/issues/4803 | [] | null | 4,803 | false |
`with_format` behavior is inconsistent on different datasets | ## Describe the bug
I found a case where `with_format` does not transform the dataset to the requested format.
## Steps to reproduce the bug
Run:
```python
from transformers import AutoTokenizer, AutoFeatureExtractor
from datasets import load_dataset
raw = load_dataset("glue", "sst2", split="train")
raw =... | https://github.com/huggingface/datasets/issues/4802 | [
"Hi! You can get a `torch.Tensor` if you do the following:\r\n```python\r\nraw = load_dataset(\"beans\", split=\"train\")\r\nraw = raw.select(range(100))\r\n\r\npreprocessor = AutoFeatureExtractor.from_pretrained(\"nateraw/vit-base-beans\")\r\n\r\nfrom datasets import Array3D\r\nfeatures = raw.features.copy()\r\nfe... | null | 4,802 | false |
Fix fine classes in trec dataset | This PR:
- replaces the fine labels, so that there are 50 instead of 47
- once more labels are added, all they (fine and coarse) have been re-ordered, so that they align with the order in: https://cogcomp.seas.upenn.edu/Data/QA/QC/definition.html
- the feature names have been fixed: `fine_label` instead of `label-fi... | https://github.com/huggingface/datasets/pull/4801 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4801",
"html_url": "https://github.com/huggingface/datasets/pull/4801",
"diff_url": "https://github.com/huggingface/datasets/pull/4801.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4801.patch",
"merged_at": "2022-08-22T16:14... | 4,801 | true |
support LargeListArray in pyarrow | ```python
import numpy as np
import datasets
a = np.zeros((5000000, 768))
res = datasets.Dataset.from_dict({'embedding': a})
'''
File '/home/wenjiaxin/anaconda3/envs/data/lib/python3.8/site-packages/datasets/arrow_writer.py', line 178, in __arrow_array__
out = numpy_to_pyarrow_listarray(data)
File "/h... | https://github.com/huggingface/datasets/pull/4800 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4800). All of your documentation changes will be reflected on that endpoint.",
"Hi, thanks for working on this! Can you run `make style` at the repo root to fix the code quality error in CI and add a test?",
"Hi, I have fixed... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4800",
"html_url": "https://github.com/huggingface/datasets/pull/4800",
"diff_url": "https://github.com/huggingface/datasets/pull/4800.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4800.patch",
"merged_at": null
} | 4,800 | true |
video dataset loader/parser | you know how you can [use `load_dataset` with any arbitrary csv file](https://huggingface.co/docs/datasets/loading#csv)? and you can also [use it to load a local image dataset](https://huggingface.co/docs/datasets/image_load#local-files)?
could you please add functionality to load a video dataset? it would be really... | https://github.com/huggingface/datasets/issues/4799 | [
"Hi! We've just started discussing the video support in `datasets` (decoding backends, video feature type, etc.), so I believe we should have something tangible by the end of this year.\r\n\r\nAlso, if you have additional video features in mind that you would like to see, feel free to let us know",
"Coool thanks ... | null | 4,799 | false |
Shard generator | Hi everyone! I was using Hugging Face datasets to process some very large datasets and found that it would be quite handy to have a feature that will allow to "split" these large datasets into chunks with equal size. Even better - be able to run through these chunks one by one in simple and convenient way. So I decided... | https://github.com/huggingface/datasets/pull/4798 | [
"Hi, thanks!\r\n\r\n> I was using Hugging Face datasets to process some very large datasets and found that it would be quite handy to have a feature that will allow to \"split\" these large datasets into chunks with equal size\r\n\r\n`map`, the method we use for processing in `datasets`, already does that if `batch... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4798",
"html_url": "https://github.com/huggingface/datasets/pull/4798",
"diff_url": "https://github.com/huggingface/datasets/pull/4798.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4798.patch",
"merged_at": null
} | 4,798 | true |
Torgo dataset creation | null | https://github.com/huggingface/datasets/pull/4797 | [
"Hi @YingLi001, thanks for your proposal to add this dataset.\r\n\r\nHowever, now we add datasets directly to the Hub (instead of our GitHub repository). You have the instructions in our docs: \r\n- [Create a dataset loading script](https://huggingface.co/docs/datasets/dataset_script)\r\n- [Create a dataset card](h... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4797",
"html_url": "https://github.com/huggingface/datasets/pull/4797",
"diff_url": "https://github.com/huggingface/datasets/pull/4797.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4797.patch",
"merged_at": null
} | 4,797 | true |
ArrowInvalid: Could not convert <PIL.Image.Image image mode=RGB when adding image to Dataset | ## Describe the bug
When adding a Pillow image to an existing Dataset on the hub, `add_item` fails due to the Pillow image not being automatically converted into the Image feature.
## Steps to reproduce the bug
```python
from datasets import load_dataset
from PIL import Image
dataset = load_dataset("hf-inte... | https://github.com/huggingface/datasets/issues/4796 | [
"@mariosasko I'm getting a similar issue when creating a Dataset from a Pandas dataframe, like so:\r\n\r\n```\r\nfrom datasets import Dataset, Features, Image, Value\r\nimport pandas as pd\r\nimport requests\r\nimport PIL\r\n\r\n# we need to define the features ourselves\r\nfeatures = Features({\r\n 'a': Value(d... | null | 4,796 | false |
Missing MBPP splits | (@albertvillanova)
The [MBPP dataset on the Hub](https://huggingface.co/datasets/mbpp) has only a test split for both its "full" and its "sanitized" subset, while the [paper](https://arxiv.org/abs/2108.07732) states in subsection 2.1 regarding the full split:
> In the experiments described later in the paper, we hold... | https://github.com/huggingface/datasets/issues/4795 | [
"Thanks for reporting this as well, @stadlerb.\r\n\r\nI suggest waiting for the answer of the data owners... ",
"@albertvillanova The first author of the paper responded to the upstream issue:\r\n> Task IDs 11-510 are the 500 test problems. We use 90 problems (511-600) for validation and then remaining 374 for fi... | null | 4,795 | false |
Add DocVQA | ## Adding a Dataset
- **Name:** DocVQA
- **Description:** Document Visual Question Answering (DocVQA) seeks to inspire a “purpose-driven” point of view in Document Analysis and Recognition research, where the document content is extracted and used to respond to high-level tasks defined by the human consumers of this ... | https://github.com/huggingface/datasets/issues/4792 | [
"Thanks for proposing, @NielsRogge.\r\n\r\nPlease, note this dataset requires registering in their website and their Terms and Conditions state we cannot distribute their URL:\r\n```\r\n1. You will NOT distribute the download URLs\r\n...\r\n```"
] | null | 4,792 | false |
Dataset Viewer issue for Team-PIXEL/rendered-wikipedia-english | ### Link
https://huggingface.co/datasets/Team-PIXEL/rendered-wikipedia-english/viewer/rendered-wikipedia-en/train
### Description
The dataset can be loaded fine but the viewer shows this error:
```
Server Error
Status code: 400
Exception: Status400Error
Message: The dataset does not exist.
```
... | https://github.com/huggingface/datasets/issues/4791 | [
"Thanks for reporting. It's a known issue that should be fixed soon. Meanwhile, I had to manually trigger the dataset viewer. It's OK now.\r\nNote that the extreme aspect ratio of the images generates another issue, that we're inspecting."
] | null | 4,791 | false |
Issue with fine classes in trec dataset | ## Describe the bug
According to their paper, the TREC dataset contains 2 kinds of classes:
- 6 coarse classes: TREC-6
- 50 fine classes: TREC-50
However, our implementation only has 47 (instead of 50) fine classes. The reason for this is that we only considered the last segment of the label, which is repeated fo... | https://github.com/huggingface/datasets/issues/4790 | [] | null | 4,790 | false |
Update doc upload_dataset.mdx | null | https://github.com/huggingface/datasets/pull/4789 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4789",
"html_url": "https://github.com/huggingface/datasets/pull/4789",
"diff_url": "https://github.com/huggingface/datasets/pull/4789.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4789.patch",
"merged_at": "2022-09-09T16:34... | 4,789 | true |
Fix NonMatchingChecksumError in mbpp dataset | Fix issue reported on the Hub: https://huggingface.co/datasets/mbpp/discussions/1
Fix #4787. | https://github.com/huggingface/datasets/pull/4788 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thank you for the quick response! Before noticing that you already had implemented the fix, I already had implemened my own version. I'd also suggest bumping the major version because the contents of the dataset changed, even if only... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4788",
"html_url": "https://github.com/huggingface/datasets/pull/4788",
"diff_url": "https://github.com/huggingface/datasets/pull/4788.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4788.patch",
"merged_at": "2022-08-04T17:21... | 4,788 | true |
NonMatchingChecksumError in mbpp dataset | ## Describe the bug
As reported on the Hub [Fix Checksum Mismatch](https://huggingface.co/datasets/mbpp/discussions/1), there is a `NonMatchingChecksumError` when loading mbpp dataset
## Steps to reproduce the bug
```python
ds = load_dataset("mbpp", "full")
```
## Expected results
Loading of the dataset with... | https://github.com/huggingface/datasets/issues/4787 | [] | null | 4,787 | false |
.save_to_disk('path', fs=s3) TypeError | The following code:
```python
import datasets
train_dataset, test_dataset = load_dataset("imdb", split=["train", "test"])
s3 = datasets.filesystems.S3FileSystem(key=aws_access_key_id, secret=aws_secret_access_key)
train_dataset.save_to_disk("s3://datasets/", fs=s3)
```
produces following traceback:
```she... | https://github.com/huggingface/datasets/issues/4786 | [] | null | 4,786 | false |
Require torchaudio<0.12.0 in docs | This PR adds to docs the requirement of torchaudio<0.12.0 to avoid RuntimeError.
Subsequent to PR:
- #4777 | https://github.com/huggingface/datasets/pull/4785 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4785",
"html_url": "https://github.com/huggingface/datasets/pull/4785",
"diff_url": "https://github.com/huggingface/datasets/pull/4785.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4785.patch",
"merged_at": "2022-08-03T14:52... | 4,785 | true |
Add Multiface dataset | ## Adding a Dataset
- **Name:** Multiface dataset
- **Description:** f high quality recordings of the faces of 13 identities, each captured in a multi-view capture stage performing various facial expressions. An average of 12,200 (v1 scripts) to 23,000 (v2 scripts) frames per subject with capture rate at 30 fps
- **... | https://github.com/huggingface/datasets/issues/4784 | [
"Hi @osanseviero I would like to add this dataset.",
"Hey @nandwalritik! Thanks for offering to help!\r\n\r\nThis dataset might be somewhat complex and I'm concerned about it being 65 TB, which would be quite expensive to host. @lhoestq @mariosasko I would love your input if you think it's worth adding this datas... | null | 4,784 | false |
Docs for creating a loading script for image datasets | This PR is a first draft of creating a loading script for image datasets. Feel free to let me know if there are any specificities I'm missing for this. 🙂
To do:
- [x] Document how to create different configurations. | https://github.com/huggingface/datasets/pull/4783 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"IMO it would make more sense to add a \"Create image dataset\" page with two main sections - a no-code approach with `imagefolder` + metadata (preferred way), and with a loading script (advanced). It should be clear when to choose wh... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4783",
"html_url": "https://github.com/huggingface/datasets/pull/4783",
"diff_url": "https://github.com/huggingface/datasets/pull/4783.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4783.patch",
"merged_at": "2022-09-07T19:07... | 4,783 | true |
pyarrow.lib.ArrowCapacityError: array cannot contain more than 2147483646 bytes, have 2147483648 | ## Describe the bug
Following the example in CodeParrot, I receive an array size limitation error when deduplicating larger datasets.
## Steps to reproduce the bug
```python
dataset_name = "the_pile"
ds = load_dataset(dataset_name, split="train")
ds = ds.map(preprocess, num_proc=num_workers)
uniques = set(ds.u... | https://github.com/huggingface/datasets/issues/4782 | [
"Thanks for reporting @conceptofmind.\r\n\r\nCould you please give details about your environment? \r\n```\r\n## Environment info\r\n<!-- You can run the command `datasets-cli env` and copy-and-paste its output below. -->\r\n- `datasets` version:\r\n- Platform:\r\n- Python version:\r\n- PyArrow version:\r\n```",
... | null | 4,782 | false |
Fix label renaming and add a battery of tests | This PR makes some changes to label renaming in `to_tf_dataset()`, both to fix some issues when users input something we weren't expecting, and also to make it easier to deprecate label renaming in future, if/when we want to move this special-casing logic to a function in `transformers`.
The main changes are:
- Lab... | https://github.com/huggingface/datasets/pull/4781 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Why don't we deprecate label renaming already instead ?",
"I think it'll break a lot of workflows if we deprecate it now! There isn't really a non-deprecated workflow yet - once we've added the `auto_rename_labels` option, then we ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4781",
"html_url": "https://github.com/huggingface/datasets/pull/4781",
"diff_url": "https://github.com/huggingface/datasets/pull/4781.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4781.patch",
"merged_at": "2022-09-12T11:24... | 4,781 | true |
Remove apache_beam import from module level in natural_questions dataset | Instead of importing `apache_beam` at the module level, import it in the method `_build_pcollection`.
Fix #4779. | https://github.com/huggingface/datasets/pull/4780 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4780",
"html_url": "https://github.com/huggingface/datasets/pull/4780",
"diff_url": "https://github.com/huggingface/datasets/pull/4780.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4780.patch",
"merged_at": "2022-08-02T16:03... | 4,780 | true |
Loading natural_questions requires apache_beam even with existing preprocessed data | ## Describe the bug
When loading "natural_questions", the package "apache_beam" is required:
```
ImportError: To be able to use natural_questions, you need to install the following dependency: apache_beam.
Please install it using 'pip install apache_beam' for instance'
```
This requirement is unnecessary, once ... | https://github.com/huggingface/datasets/issues/4779 | [] | null | 4,779 | false |
Update local loading script docs | This PR clarifies the local loading script section to include how to load a dataset after you've modified the local loading script (closes #4732). | https://github.com/huggingface/datasets/pull/4778 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4778). All of your documentation changes will be reflected on that endpoint.",
"I would rather have a section in the docs that explains how to modify the script of an existing dataset (`inspect_dataset` + modification + `load_d... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4778",
"html_url": "https://github.com/huggingface/datasets/pull/4778",
"diff_url": "https://github.com/huggingface/datasets/pull/4778.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4778.patch",
"merged_at": "2022-08-23T16:32... | 4,778 | true |
Require torchaudio<0.12.0 to avoid RuntimeError | Related to:
- https://github.com/huggingface/transformers/issues/18379
Fix partially #4776. | https://github.com/huggingface/datasets/pull/4777 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4777",
"html_url": "https://github.com/huggingface/datasets/pull/4777",
"diff_url": "https://github.com/huggingface/datasets/pull/4777.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4777.patch",
"merged_at": "2022-08-02T17:21... | 4,777 | true |
RuntimeError when using torchaudio 0.12.0 to load MP3 audio file | Current version of `torchaudio` (0.12.0) raises a RuntimeError when trying to use `sox_io` backend but non-Python dependency `sox` is not installed:
https://github.com/pytorch/audio/blob/2e1388401c434011e9f044b40bc8374f2ddfc414/torchaudio/backend/sox_io_backend.py#L21-L29
```python
def _fail_load(
filepath: str... | https://github.com/huggingface/datasets/issues/4776 | [
"Requiring torchaudio<0.12.0 isn't really a viable solution because that implies torch<0.12.0 which means no sm_86 CUDA support which means no RTX 3090 support in PyTorch.\r\n\r\nBut in my case, the error only occurs if `_fallback_load` resolves to `_fail_load` inside torchaudio 0.12.0 which is only the case if FFM... | null | 4,776 | false |
Streaming not supported in Theivaprakasham/wildreceipt | ### Link
_No response_
### Description
_No response_
### Owner
_No response_ | https://github.com/huggingface/datasets/issues/4775 | [
"Thanks for reporting @NitishkKarra.\r\n\r\nThe root source of the issue is that streaming mode is not supported out-of-the-box for that dataset, because it contains a TAR file.\r\n\r\nWe have opened a discussion in the corresponding Hub dataset page, pointing out this issue: https://huggingface.co/datasets/Theivap... | null | 4,775 | false |
Training hangs at the end of epoch, with set_transform/with_transform+multiple workers | ## Describe the bug
I use load_dataset() (I tried with [wiki](https://huggingface.co/datasets/wikipedia) and my own json data) and use set_transform/with_transform for preprocessing. But it hangs at the end of the 1st epoch if dataloader_num_workers>=1. No problem with single worker.
## Steps to reproduce the bu... | https://github.com/huggingface/datasets/issues/4774 | [] | null | 4,774 | false |
Document loading from relative path | This PR describes loading a dataset from the Hub by specifying a relative path in `data_dir` or `data_files` in `load_dataset` (see #4757). | https://github.com/huggingface/datasets/pull/4773 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks for the feedback!\r\n\r\nI agree that adding it to `load_hub.mdx` is probably a bit too specific, especially for beginners reading the tutorials. Since this clarification is closely related to loading from the Hub (the only di... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4773",
"html_url": "https://github.com/huggingface/datasets/pull/4773",
"diff_url": "https://github.com/huggingface/datasets/pull/4773.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4773.patch",
"merged_at": "2022-08-25T18:34... | 4,773 | true |
AssertionError when using label_cols in to_tf_dataset | ## Describe the bug
An incorrect `AssertionError` is raised when using `label_cols` in `to_tf_dataset` and the label's key name is `label`.
The assertion is in this line:
https://github.com/huggingface/datasets/blob/2.4.0/src/datasets/arrow_dataset.py#L475
## Steps to reproduce the bug
```python
from datasets... | https://github.com/huggingface/datasets/issues/4772 | [
"cc @Rocketknight1 ",
"Hi @lehrig, this is caused by the data collator renaming \"label\" to \"labels\". If you set `label_cols=[\"labels\"]` in the call it will work correctly. However, I agree that the cause of the bug is not obvious, so I'll see if I can make a PR to clarify things when the collator renames co... | null | 4,772 | false |
Remove dummy data generation docs | This PR removes instructions to generate dummy data since that is no longer necessary for datasets that are uploaded to the Hub instead of our GitHub repo.
Close #4744 | https://github.com/huggingface/datasets/pull/4771 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4771",
"html_url": "https://github.com/huggingface/datasets/pull/4771",
"diff_url": "https://github.com/huggingface/datasets/pull/4771.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4771.patch",
"merged_at": "2022-08-02T23:50... | 4,771 | true |
fix typo | By defaul -> By default | https://github.com/huggingface/datasets/pull/4770 | [
"good catch thanks ! Can you check if the same typo is also present in `add_elasticsearch_index` ? It has a very similar signature",
"> good catch thanks ! Can you check if the same typo is also present in `add_elasticsearch_index` ? It has a very similar signature\r\n\r\nfixed"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4770",
"html_url": "https://github.com/huggingface/datasets/pull/4770",
"diff_url": "https://github.com/huggingface/datasets/pull/4770.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4770.patch",
"merged_at": "2022-07-29T16:02... | 4,770 | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.