title stringlengths 1 290 | body stringlengths 0 228k โ | html_url stringlengths 46 51 | comments list | pull_request dict | number int64 1 5.59k | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|
Make HANS dataset streamable | Fix #4133 | https://github.com/huggingface/datasets/pull/4155 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4155",
"html_url": "https://github.com/huggingface/datasets/pull/4155",
"diff_url": "https://github.com/huggingface/datasets/pull/4155.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4155.patch",
"merged_at": "2022-04-13T11:57... | 4,155 | true |
Generate tasks.json taxonomy from `huggingface_hub` | null | https://github.com/huggingface/datasets/pull/4154 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Ok recomputed the json file, this should be ready to review now! @lhoestq ",
"Note: the generated JSON from `hf/hub-docs` can be found in the output of a GitHub Action run on that repo, for instance in https://github.com/huggingfac... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4154",
"html_url": "https://github.com/huggingface/datasets/pull/4154",
"diff_url": "https://github.com/huggingface/datasets/pull/4154.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4154.patch",
"merged_at": "2022-04-14T10:26... | 4,154 | true |
Adding Text-based NP Enrichment (TNE) dataset | Added the [TNE](https://github.com/yanaiela/TNE) dataset to the library | https://github.com/huggingface/datasets/pull/4153 | [
"Hey @lhoestq, can you please have a look? ๐",
"Great, thanks again @lhoestq! I think we're good to go now",
"Done"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4153",
"html_url": "https://github.com/huggingface/datasets/pull/4153",
"diff_url": "https://github.com/huggingface/datasets/pull/4153.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4153.patch",
"merged_at": "2022-05-03T14:05... | 4,153 | true |
ArrayND error in pyarrow 5 | As found in https://github.com/huggingface/datasets/pull/3903, The ArrayND features fail on pyarrow 5:
```python
import pyarrow as pa
from datasets import Array2D
from datasets.table import cast_array_to_feature
arr = pa.array([[[0]]])
feature_type = Array2D(shape=(1, 1), dtype="int64")
cast_array_to_feature(a... | https://github.com/huggingface/datasets/issues/4152 | [
"Where do we bump the required pyarrow version? Any inputs on how I fix this issue? ",
"We need to bump it in `setup.py` as well as update some CI job to use pyarrow 6 instead of 5 in `.circleci/config.yaml` and `.github/workflows/benchmarks.yaml`"
] | null | 4,152 | false |
Add missing label for emotion description | null | https://github.com/huggingface/datasets/pull/4151 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4151",
"html_url": "https://github.com/huggingface/datasets/pull/4151",
"diff_url": "https://github.com/huggingface/datasets/pull/4151.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4151.patch",
"merged_at": "2022-04-12T13:58... | 4,151 | true |
Inconsistent splits generation for datasets without loading script (packaged dataset puts everything into a single split) | ## Describe the bug
Splits for dataset loaders without scripts are prepared inconsistently. I think it might be confusing for users.
## Steps to reproduce the bug
* If you load a packaged datasets from Hub, it infers splits from directory structure / filenames (check out the data [here](https://huggingface.co/data... | https://github.com/huggingface/datasets/issues/4150 | [] | null | 4,150 | false |
load_dataset for winoground returning decoding error | ## Describe the bug
I am trying to use datasets to load winoground and I'm getting a JSON decoding error.
## Steps to reproduce the bug
```python
from datasets import load_dataset
token = 'hf_XXXXX' # my HF access token
datasets = load_dataset('facebook/winoground', use_auth_token=token)
```
## Expected res... | https://github.com/huggingface/datasets/issues/4149 | [
"I thought I had fixed it with this after some helpful hints from @severo\r\n```python\r\nimport datasets \r\ntoken = 'hf_XXXXX'\r\ndataset = datasets.load_dataset(\r\n 'facebook/winoground', \r\n name='facebook--winoground', \r\n split='train', \r\n streaming=True,\r\n use_auth_token=token,\r\n)\r\n... | null | 4,149 | false |
fix confusing bleu metric example | **Is your feature request related to a problem? Please describe.**
I would like to see the example in "Metric Card for BLEU" changed.
The 0th element in the predictions list is not closed in square brackets, and the 1st list is missing a comma.
The BLEU score are calculated correctly, but it is difficult to understa... | https://github.com/huggingface/datasets/issues/4148 | [] | null | 4,148 | false |
Adjust path to datasets tutorial in How-To | The link in the How-To overview page to the Datasets tutorials is currently broken. This is just a small adjustment to make it match the format used in https://github.com/huggingface/datasets/blob/master/docs/source/tutorial.md.
(Edit to add: The link in the PR deployment (https://moon-ci-docs.huggingface.co/docs/da... | https://github.com/huggingface/datasets/pull/4147 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4147",
"html_url": "https://github.com/huggingface/datasets/pull/4147",
"diff_url": "https://github.com/huggingface/datasets/pull/4147.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4147.patch",
"merged_at": "2022-04-12T08:26... | 4,147 | true |
SAMSum dataset viewer not working | ## Dataset viewer issue for '*name of the dataset*'
**Link:** *link to the dataset viewer page*
*short description of the issue*
Am I the one who added this dataset ? Yes-No
| https://github.com/huggingface/datasets/issues/4146 | [
"https://huggingface.co/datasets/samsum\r\n\r\n```\r\nStatus code: 400\r\nException: ValueError\r\nMessage: Cannot seek streaming HTTP file\r\n```",
"Currently, only the datasets that can be streamed support the dataset viewer. Maybe @lhoestq @albertvillanova or @mariosasko could give more details abo... | null | 4,146 | false |
Redirect TIMIT download from LDC | LDC data is protected under US copyright laws and under various legal agreements between the Linguistic Data Consortium/the University of Pennsylvania and data providers which prohibit redistribution of that data by anyone other than LDC. Similarly, LDC's membership agreements, non-member user agreement and various cor... | https://github.com/huggingface/datasets/pull/4145 | [
"CI is failing because some tags are outdated, but they're fixed in #4067 ",
"_The documentation is not available anymore as the PR was closed or merged._",
"We may do a release pretty soon (today ?), let me know if it's fine to include it in the new release",
"Fine to include this change!"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4145",
"html_url": "https://github.com/huggingface/datasets/pull/4145",
"diff_url": "https://github.com/huggingface/datasets/pull/4145.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4145.patch",
"merged_at": "2022-04-13T15:33... | 4,145 | true |
Fix splits in local packaged modules, local datasets without script and hub datasets without script | fixes #4150
I suggest to infer splits structure from files when `data_dir` is passed with `get_patterns_locally`, analogous to what's done in `LocalDatasetModuleFactoryWithoutScript` with `self.path`, instead of generating files with `data_dir/**` patterns and putting them all into a single default (train) split.
... | https://github.com/huggingface/datasets/pull/4144 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks !\r\nI'm in favor of this change, even though it's a breaking change:\r\n\r\nif you had a dataset\r\n```\r\ndata/\r\n train.csv\r\n test.csv\r\n```\r\n\r\nthen running this code would now return both train and test splits:\r... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4144",
"html_url": "https://github.com/huggingface/datasets/pull/4144",
"diff_url": "https://github.com/huggingface/datasets/pull/4144.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4144.patch",
"merged_at": "2022-04-28T21:02... | 4,144 | true |
Unable to download `Wikepedia` 20220301.en version | ## Describe the bug
Unable to download `Wikepedia` dataset, 20220301.en version
## Steps to reproduce the bug
```python
!pip install apache_beam mwparserfromhell
dataset_wikipedia = load_dataset("wikipedia", "20220301.en")
```
## Actual results
```
ValueError: BuilderConfig 20220301.en not found.
Avail... | https://github.com/huggingface/datasets/issues/4143 | [
"Hi! We've recently updated the Wikipedia script, so these changes are only available on master and can be fetched as follows:\r\n```python\r\ndataset_wikipedia = load_dataset(\"wikipedia\", \"20220301.en\", revision=\"master\")\r\n```",
"Hi, how can I load the previous \"20200501.en\" version of wikipedia which ... | null | 4,143 | false |
Add ObjectFolder 2.0 dataset | ## Adding a Dataset
- **Name:** ObjectFolder 2.0
- **Description:** ObjectFolder 2.0 is a dataset of 1,000 objects in the form of implicit representations. It contains 1,000 Object Files each containing the complete multisensory profile for an object instance.
- **Paper:** [*link to the dataset paper if available*](... | https://github.com/huggingface/datasets/issues/4142 | [
"Datasets are not tracked in this repository anymore."
] | null | 4,142 | false |
Why is the dataset not visible under the dataset preview section? | ## Dataset viewer issue for '*name of the dataset*'
**Link:** *link to the dataset viewer page*
*short description of the issue*
Am I the one who added this dataset ? Yes-No
| https://github.com/huggingface/datasets/issues/4141 | [] | null | 4,141 | false |
Error loading arxiv data set | ## Describe the bug
A clear and concise description of what the bug is.
I met the error below when loading arxiv dataset via `nlp.load_dataset('scientific_papers', 'arxiv',)`.
```
Traceback (most recent call last):
File "scripts/summarization.py", line 354, in <module>
main(args)
File "scripts/summari... | https://github.com/huggingface/datasets/issues/4140 | [
"Hi! I think this error may be related to using an older version of the library. I was able to load the dataset without any issues using the latest version of `datasets`. Can you upgrade to the latest version of `datasets` and try again? :)",
"Hi! As @stevhliu suggested, to fix the issue, update the lib to the n... | null | 4,140 | false |
Dataset viewer issue for Winoground | ## Dataset viewer issue for 'Winoground'
**Link:** [*link to the dataset viewer page*](https://huggingface.co/datasets/facebook/winoground/viewer/facebook--winoground/train)
*short description of the issue*
Getting 401, message='Unauthorized'
The dataset is subject to authorization, but I can access the files f... | https://github.com/huggingface/datasets/issues/4139 | [
"related (same dataset): https://github.com/huggingface/datasets/issues/4149. But the issue is different. Looking at it",
"I thought this issue was related to the error I was seeing, but upon consideration I'd think the dataset viewer would return a 500 (unable to create the split like me) or a 404 (unable to loa... | null | 4,139 | false |
Incorrect Russian filenames encoding after extraction by datasets.DownloadManager.download_and_extract() | ## Dataset viewer issue for 'MalakhovIlya/RuREBus'
**Link:** https://huggingface.co/datasets/MalakhovIlya/RuREBus
**Description**
Using os.walk(topdown=False) in DatasetBuilder causes following error:
Status code: 400
Exception: TypeError
Message: xwalk() got an unexpected keyword argument 'topdow... | https://github.com/huggingface/datasets/issues/4138 | [
"To reproduce:\r\n\r\n```python\r\n>>> import datasets\r\n>>> datasets.get_dataset_split_names('MalakhovIlya/RuREBus', config_name='raw_txt')\r\nTraceback (most recent call last):\r\n File \"/home/slesage/hf/datasets-preview-backend/.venv/lib/python3.9/site-packages/datasets/inspect.py\", line 280, in get_dataset_... | null | 4,138 | false |
Add single dataset citations for TweetEval | This PR adds single data citations as per request of the original creators of the TweetEval dataset.
This is a recent email from the creator:
> Could I ask you a favor? Would you be able to add at the end of the README the citations of the single datasets as well? You can just copy our readme maybe? https://githu... | https://github.com/huggingface/datasets/pull/4137 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The `test_dataset_cards` method is failing with the error:\r\n\r\n```\r\nif error_messages:\r\n> raise ValueError(\"\\n\".join(error_messages))\r\nE ValueError: The following issues have been found in the dataset ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4137",
"html_url": "https://github.com/huggingface/datasets/pull/4137",
"diff_url": "https://github.com/huggingface/datasets/pull/4137.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4137.patch",
"merged_at": "2022-04-12T07:51... | 4,137 | true |
Support streaming xtreme dataset for PAN-X config | Support streaming xtreme dataset for PAN-X config. | https://github.com/huggingface/datasets/pull/4135 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4135",
"html_url": "https://github.com/huggingface/datasets/pull/4135",
"diff_url": "https://github.com/huggingface/datasets/pull/4135.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4135.patch",
"merged_at": "2022-04-11T06:54... | 4,135 | true |
ELI5 supporting documents | if i am using dense search to create supporting documents for eli5 how much time it will take bcz i read somewhere that it takes about 18 hrs?? | https://github.com/huggingface/datasets/issues/4134 | [
"Hi ! Please post your question on the [forum](https://discuss.huggingface.co/), more people will be able to help you there ;)"
] | null | 4,134 | false |
HANS dataset preview broken | ## Dataset viewer issue for '*hans*'
**Link:** [https://huggingface.co/datasets/hans](https://huggingface.co/datasets/hans)
HANS dataset preview is broken with error 400
Am I the one who added this dataset ? No
| https://github.com/huggingface/datasets/issues/4133 | [
"The dataset cannot be loaded, be it in normal or streaming mode.\r\n\r\n```python\r\n>>> import datasets\r\n>>> dataset=datasets.load_dataset(\"hans\", split=\"train\", streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/sles... | null | 4,133 | false |
Support streaming xtreme dataset for PAWS-X config | Support streaming xtreme dataset for PAWS-X config. | https://github.com/huggingface/datasets/pull/4132 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4132",
"html_url": "https://github.com/huggingface/datasets/pull/4132",
"diff_url": "https://github.com/huggingface/datasets/pull/4132.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4132.patch",
"merged_at": "2022-04-08T21:02... | 4,132 | true |
Support streaming xtreme dataset for udpos config | Support streaming xtreme dataset for udpos config. | https://github.com/huggingface/datasets/pull/4131 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4131",
"html_url": "https://github.com/huggingface/datasets/pull/4131",
"diff_url": "https://github.com/huggingface/datasets/pull/4131.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4131.patch",
"merged_at": "2022-04-08T16:28... | 4,131 | true |
Add SBU Captions Photo Dataset | null | https://github.com/huggingface/datasets/pull/4130 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4130",
"html_url": "https://github.com/huggingface/datasets/pull/4130",
"diff_url": "https://github.com/huggingface/datasets/pull/4130.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4130.patch",
"merged_at": "2022-04-12T10:41... | 4,130 | true |
dataset metadata for reproducibility | When pulling a dataset from the hub, it would be useful to have some metadata about the specific dataset and version that is used. The metadata could then be passed to the `Trainer` which could then be saved to a model card. This is useful for people who run many experiments on different versions (commits/branches) of ... | https://github.com/huggingface/datasets/issues/4129 | [] | null | 4,129 | false |
More robust `cast_to_python_objects` in `TypedSequence` | Adds a fallback to run an expensive version of `cast_to_python_objects` which exhaustively checks entire lists to avoid the `ArrowInvalid: Could not convert` error in `TypedSequence`. Currently, this error can happen in situations where only some images are decoded in `map`, in which case `cast_to_python_objects` fails... | https://github.com/huggingface/datasets/pull/4128 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4128",
"html_url": "https://github.com/huggingface/datasets/pull/4128",
"diff_url": "https://github.com/huggingface/datasets/pull/4128.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4128.patch",
"merged_at": "2022-04-13T14:01... | 4,128 | true |
Add configs with processed data in medical_dialog dataset | There exist processed data files that do not require parsing the raw data files (which can take long time).
Fix #4122. | https://github.com/huggingface/datasets/pull/4127 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4127",
"html_url": "https://github.com/huggingface/datasets/pull/4127",
"diff_url": "https://github.com/huggingface/datasets/pull/4127.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4127.patch",
"merged_at": "2022-04-08T16:20... | 4,127 | true |
dataset viewer issue for common_voice | ## Dataset viewer issue for 'common_voice'
**Link:** https://huggingface.co/datasets/common_voice
Server Error
Status code: 400
Exception: TypeError
Message: __init__() got an unexpected keyword argument 'audio_column'
Am I the one who added this dataset ? No
| https://github.com/huggingface/datasets/issues/4126 | [
"Yes, it's a known issue, and we expect to fix it soon.",
"Fixed.\r\n\r\n<img width=\"1393\" alt=\"Capture dโeฬcran 2022-04-25 aฬ 15 42 05\" src=\"https://user-images.githubusercontent.com/1676121/165101176-d729d85b-efff-45a8-bad1-b69223edba5f.png\">\r\n"
] | null | 4,126 | false |
BIG-bench | This PR adds all BIG-bench json tasks to huggingface/datasets. | https://github.com/huggingface/datasets/pull/4125 | [
"> It looks like the CI is failing on windows because our windows CI is unable to clone the bigbench repository (maybe it has to do with filenames that are longer than 256 characters, which windows don't like). Could the smaller installation of bigbench via pip solve this issue ?\r\n> Otherwise we can see how to re... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4125",
"html_url": "https://github.com/huggingface/datasets/pull/4125",
"diff_url": "https://github.com/huggingface/datasets/pull/4125.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4125.patch",
"merged_at": "2022-06-08T17:32... | 4,125 | true |
Image decoding often fails when transforming Image datasets | ## Describe the bug
When transforming/modifying images in an image dataset using the `map` function the PIL images often fail to decode in time for the image transforms, causing errors.
Using a debugger it is easy to see what the problem is, the Image decode invocation does not take place and the resulting image pa... | https://github.com/huggingface/datasets/issues/4124 | [
"A quick hack I have found is that we can call the image first before running the transforms and it makes sure the image is decoded before being passed on.\r\n\r\nFor this I just needed to add `example['img'] = example['img']` to the top of my `generate_flipped_data` function, defined above, so that image decode in... | null | 4,124 | false |
Building C4 takes forever | ## Describe the bug
C4-en is a 300 GB dataset. However, when I try to download it through the hub it takes over _six hours_ to generate the train/test split from the downloaded files. This is an absurd amount of time and an unnecessary waste of resources.
## Steps to reproduce the bug
```python
c4 = datasets.load... | https://github.com/huggingface/datasets/issues/4123 | [
"Hi @StellaAthena, thanks for reporting.\r\n\r\nPlease note, that our `datasets` library performs several operations in order to load a dataset, among them:\r\n- it downloads all the required files: for C4 \"en\", 378.69 GB of JSON GZIPped files\r\n- it parses their content to generate the dataset\r\n- it caches th... | null | 4,123 | false |
medical_dialog zh has very slow _generate_examples | ## Describe the bug
After downloading the files from Google Drive, `load_dataset("medical_dialog", "zh", data_dir="./")` takes an unreasonable amount of time. Generating the train/test split for 33% of the dataset takes over 4.5 hours.
## Steps to reproduce the bug
The easiest way I've found to download files from... | https://github.com/huggingface/datasets/issues/4122 | [
"Hi @nbroad1881, thanks for reporting.\r\n\r\nLet me have a look to try to improve its performance. ",
"Thanks @nbroad1881 for reporting! I don't recall it taking so long. I will also have a look at this. \r\n@albertvillanova please let me know if I am doing something unnecessary or time consuming.",
"Hi @nbro... | null | 4,122 | false |
datasets.load_metric can not load a local metirc | ## Describe the bug
No matter how I hard try to tell load_metric that I want to load a local metric file, it still continues to fetch things on the Internet. And unfortunately it says 'ConnectionError: Couldn't reach'. However I can download this file without connectionerror and tell load_metric its local directory. A... | https://github.com/huggingface/datasets/issues/4121 | [
"Hello, could you tell me how this issue can be fixed? I'm coming across the same issue."
] | null | 4,121 | false |
Representing dictionaries (json) objects as features | In the process of adding a new dataset to the hub, I stumbled upon the inability to represent dictionaries that contain different key names, unknown in advance (and may differ between samples), original asked in the [forum](https://discuss.huggingface.co/t/representing-nested-dictionary-with-different-keys/16442).
F... | https://github.com/huggingface/datasets/issues/4120 | [] | null | 4,120 | false |
Hotfix failing CI tests on Windows | This PR makes a hotfix for our CI Windows tests: https://app.circleci.com/pipelines/github/huggingface/datasets/11092/workflows/9cfdb1dd-0fec-4fe0-8122-5f533192ebdc/jobs/67414
Fix #4118
I guess this issue is related to this PR:
- huggingface/huggingface_hub#815 | https://github.com/huggingface/datasets/pull/4119 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4119",
"html_url": "https://github.com/huggingface/datasets/pull/4119",
"diff_url": "https://github.com/huggingface/datasets/pull/4119.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4119.patch",
"merged_at": "2022-04-07T07:57... | 4,119 | true |
Failing CI tests on Windows | ## Describe the bug
Our CI Windows tests are failing from yesterday: https://app.circleci.com/pipelines/github/huggingface/datasets/11092/workflows/9cfdb1dd-0fec-4fe0-8122-5f533192ebdc/jobs/67414
| https://github.com/huggingface/datasets/issues/4118 | [] | null | 4,118 | false |
AttributeError: module 'huggingface_hub' has no attribute 'hf_api' | ## Describe the bug
Could you help me please. I got this following error.
AttributeError: module 'huggingface_hub' has no attribute 'hf_api'
## Steps to reproduce the bug
when I imported the datasets
# Sample code to reproduce the bug
from datasets import list_datasets, load_dataset, list_metrics, load_metr... | https://github.com/huggingface/datasets/issues/4117 | [
"Hi @arymbe, thanks for reporting.\r\n\r\nUnfortunately, I'm not able to reproduce your problem.\r\n\r\nCould you please write the complete stack trace? That way we will be able to see which package originates the exception.",
"Hello, thank you for your fast replied. this is the complete error that I got\r\n\r\n-... | null | 4,117 | false |
Pretty print dataset info files | Adds indentation to the `dataset_infos.json` file when saving for nicer diffs.
(suggested by @julien-c)
This PR also updates the info files of the GH datasets. Note that this change adds more than **10 MB** to the repo size (the total file size before the change: 29.672298 MB, after: 41.666475 MB), so I'm not sur... | https://github.com/huggingface/datasets/pull/4116 | [
"maybe just do it from now on no? (i.e. not for existing `dataset_infos.json` files)",
"_The documentation is not available anymore as the PR was closed or merged._",
"> maybe just do it from now on no? (i.e. not for existing dataset_infos.json files)\r\n\r\nYes, or do this only for datasets created with `push_... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4116",
"html_url": "https://github.com/huggingface/datasets/pull/4116",
"diff_url": "https://github.com/huggingface/datasets/pull/4116.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4116.patch",
"merged_at": "2022-04-08T11:21... | 4,116 | true |
ImageFolder add option to ignore some folders like '.ipynb_checkpoints' | **Is your feature request related to a problem? Please describe.**
I sometimes like to peek at the dataset images from jupyterlab. thus '.ipynb_checkpoints' folder appears where my dataset is and (just realized) leads to accidental duplicate image additions. I think this is an easy enough thing to miss especially if t... | https://github.com/huggingface/datasets/issues/4115 | [
"Maybe it would be nice to ignore private dirs like this one (ones starting with `.`) by default. \r\n\r\nCC @mariosasko ",
"Maybe we can add a `ignore_hidden_files` flag to the builder configs of our packaged loaders (to be consistent across all of them), wdyt @lhoestq @albertvillanova? ",
"I think they should... | null | 4,115 | false |
Allow downloading just some columns of a dataset | **Is your feature request related to a problem? Please describe.**
Some people are interested in doing label analysis of a CV dataset without downloading all the images. Downloading the whole dataset does not always makes sense for this kind of use case
**Describe the solution you'd like**
Be able to just download... | https://github.com/huggingface/datasets/issues/4114 | [
"In the general case you canโt always reduce the quantity of data to download, since you canโt parse CSV or JSON data without downloading the whole files right ? ^^ However we could explore this case-by-case I guess",
"Actually for csv pandas has `usecols` which allows loading a subset of columns in a more effici... | null | 4,114 | false |
Multiprocessing with FileLock fails in python 3.9 | On python 3.9, this code hangs:
```python
from multiprocessing import Pool
from filelock import FileLock
def run(i):
print(f"got the lock in multi process [{i}]")
with FileLock("tmp.lock"):
with Pool(2) as pool:
pool.map(run, range(2))
```
This is because the subprocesses try to ac... | https://github.com/huggingface/datasets/issues/4113 | [
"Closing this one because it must be used this way actually:\r\n```python\r\ndef main():\r\n with FileLock(\"tmp.lock\"):\r\n with Pool(2) as pool:\r\n pool.map(run, range(2))\r\n\r\nif __name__ == \"__main__\":\r\n main()\r\n```"
] | null | 4,113 | false |
ImageFolder with Grayscale images dataset | Hi, I'm facing a problem with a grayscale images dataset I have uploaded [here](https://huggingface.co/datasets/ChainYo/rvl-cdip) (RVL-CDIP)
I'm getting an error while I want to use images for training a model with PyTorch DataLoader. Here is the full traceback:
```bash
AttributeError: Caught AttributeError in D... | https://github.com/huggingface/datasets/issues/4112 | [
"Hi! Replacing:\r\n```python\r\ntransformed_dataset = dataset.with_transform(transforms)\r\ntransformed_dataset.set_format(type=\"torch\", device=\"cuda\")\r\n```\r\n\r\nwith:\r\n```python\r\ndef transform_func(examples):\r\n examples[\"image\"] = [transforms(img).to(\"cuda\") for img in examples[\"image\"]]\r\n... | null | 4,112 | false |
Update security policy | null | https://github.com/huggingface/datasets/pull/4111 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4111",
"html_url": "https://github.com/huggingface/datasets/pull/4111",
"diff_url": "https://github.com/huggingface/datasets/pull/4111.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4111.patch",
"merged_at": "2022-04-07T09:40... | 4,111 | true |
Matthews Correlation Metric Card | null | https://github.com/huggingface/datasets/pull/4110 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4110",
"html_url": "https://github.com/huggingface/datasets/pull/4110",
"diff_url": "https://github.com/huggingface/datasets/pull/4110.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4110.patch",
"merged_at": "2022-05-03T13:36... | 4,110 | true |
Add Spearmanr Metric Card | null | https://github.com/huggingface/datasets/pull/4109 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"changes made! @lhoestq let me know what you think ",
"The CI fail is unrelated to this PR and fixed on master, feel free to merge :)"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4109",
"html_url": "https://github.com/huggingface/datasets/pull/4109",
"diff_url": "https://github.com/huggingface/datasets/pull/4109.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4109.patch",
"merged_at": "2022-05-03T16:43... | 4,109 | true |
Perplexity Speedup | This PR makes necessary changes to perplexity such that:
- it runs much faster (via batching)
- it throws an error when input is empty, or when input is one word without <BOS> token
- it adds the option to add a <BOS> token
Issues:
- The values returned are extremely high, and I'm worried they aren't correct. Ev... | https://github.com/huggingface/datasets/pull/4108 | [
"WRT the high values, can you add some unit tests with some [string, model] pairs and their resulting perplexity code, and @TristanThrush can run the same pairs through his version of the code?",
"_The documentation is not available anymore as the PR was closed or merged._",
"I thought that the perplexity metri... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4108",
"html_url": "https://github.com/huggingface/datasets/pull/4108",
"diff_url": "https://github.com/huggingface/datasets/pull/4108.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4108.patch",
"merged_at": "2022-04-20T12:54... | 4,108 | true |
Unable to view the dataset and loading the same dataset throws the error - ArrowInvalid: Exceeded maximum rows | ## Dataset viewer issue - -ArrowInvalid: Exceeded maximum rows
**Link:** *https://huggingface.co/datasets/Pavithree/explainLikeImFive*
*This is the subset of original eli5 dataset https://huggingface.co/datasets/vblagoje/lfqa. I just filtered the data samples which belongs to one particular subreddit thread. How... | https://github.com/huggingface/datasets/issues/4107 | [
"Thanks for reporting. I'm looking at it",
" It's not related to the dataset viewer in itself. I can replicate the error with:\r\n\r\n```\r\n>>> import datasets as ds\r\n>>> d = ds.load_dataset('Pavithree/explainLikeImFive')\r\nUsing custom data configuration Pavithree--explainLikeImFive-b68b6d8112cd8a51\r\nDown... | null | 4,107 | false |
Support huggingface_hub 0.5 | Following https://github.com/huggingface/datasets/issues/4105
`huggingface_hub` deprecated some parameters in `HfApi` in 0.5. This PR updates all the calls to HfApi to remove all the deprecations, <s>and I set the `hugginface_hub` requirement to `>=0.5.0`</s>
cc @adrinjalali @LysandreJik | https://github.com/huggingface/datasets/pull/4106 | [
"Looks like GH actions is not able to resolve `huggingface_hub` 0.5.0, I'm investivating",
"_The documentation is not available anymore as the PR was closed or merged._",
"I'm glad to see changes in `huggingface_hub` are simplifying code here.",
"seems to supersede #4102, feel free to close mine :)",
"maybe... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4106",
"html_url": "https://github.com/huggingface/datasets/pull/4106",
"diff_url": "https://github.com/huggingface/datasets/pull/4106.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4106.patch",
"merged_at": "2022-04-08T10:22... | 4,106 | true |
push to hub fails with huggingface-hub 0.5.0 | ## Describe the bug
`ds.push_to_hub` is failing when updating a dataset in the form "org_id/repo_id"
## Steps to reproduce the bug
```python
from datasets import load_dataset
ds = load_dataset("rubrix/news_test")
ds.push_to_hub("<your-user>/news_test", token="<your-token>")
```
## Expected results
The data... | https://github.com/huggingface/datasets/issues/4105 | [
"Hi ! Indeed there was a breaking change in `huggingface_hub` 0.5.0 in `HfApi.create_repo`, which is called here in `datasets` by passing the org name in both the `repo_id` and the `organization` arguments:\r\n\r\nhttps://github.com/huggingface/datasets/blob/2230f7f7d7fbaf102cff356f5a8f3bd1561bea43/src/datasets/arr... | null | 4,105 | false |
Add time series data - stock market | ## Adding a Time Series Dataset
- **Name:** 2min ticker data for stock market
- **Description:** 8 stocks' data collected for 1month post ukraine-russia war. 4 NSE stocks and 4 NASDAQ stocks. Along with technical indicators (additional features) as shown in below image
- **Data:** Collected by myself from investing... | https://github.com/huggingface/datasets/issues/4104 | [
"Can I use instructions present in below link for time series dataset as well? \r\nhttps://github.com/huggingface/datasets/blob/master/ADD_NEW_DATASET.md ",
"cc'ing @kashif and @NielsRogge for visibility!",
"@INF800 happy to add this dataset! I will try to set a PR by the end of the day... if you can kindly poi... | null | 4,104 | false |
Add the `GSM8K` dataset | null | https://github.com/huggingface/datasets/pull/4103 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"The CI is failing because it's outdated, but the task tags are updated on `master`, merging :)"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4103",
"html_url": "https://github.com/huggingface/datasets/pull/4103",
"diff_url": "https://github.com/huggingface/datasets/pull/4103.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4103.patch",
"merged_at": "2022-04-12T10:21... | 4,103 | true |
[hub] Fix `api.create_repo` call? | null | https://github.com/huggingface/datasets/pull/4102 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4102). All of your documentation changes will be reflected on that endpoint.",
"Closing in favor of https://github.com/huggingface/datasets/pull/4106"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4102",
"html_url": "https://github.com/huggingface/datasets/pull/4102",
"diff_url": "https://github.com/huggingface/datasets/pull/4102.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4102.patch",
"merged_at": null
} | 4,102 | true |
How can I download only the train and test split for full numbers using load_dataset()? | How can I download only the train and test split for full numbers using load_dataset()?
I do not need the extra split and it will take 40 mins just to download in Colab. I have very short time in hand. Please help. | https://github.com/huggingface/datasets/issues/4101 | [
"Hi! Can you please specify the full name of the dataset? IIRC `full_numbers` is one of the configs of the `svhn` dataset, and its generation is slow due to data being stored in binary Matlab files. Even if you specify a specific split, `datasets` downloads all of them, but we plan to fix that soon and only downloa... | null | 4,101 | false |
Improve RedCaps dataset card | This PR modifies the RedCaps card to:
* fix the formatting of the Point of Contact fields on the Hub
* speed up the image fetching logic (aligns it with the [img2dataset](https://github.com/rom1504/img2dataset) tool) and make it more robust (return None if **any** exception is thrown) | https://github.com/huggingface/datasets/pull/4100 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I find this preprocessing a bit too specific to add it as a method to `datasets` as it's only useful in the context of CV (and we support multiple modalities). However, I agree it would be great to move this code to another lib to av... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4100",
"html_url": "https://github.com/huggingface/datasets/pull/4100",
"diff_url": "https://github.com/huggingface/datasets/pull/4100.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4100.patch",
"merged_at": "2022-04-13T14:02... | 4,100 | true |
UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 213: ordinal not in range(128) | ## Describe the bug
Error "UnicodeDecodeError: 'ascii' codec can't decode byte 0xe5 in position 213: ordinal not in range(128)" is thrown when downloading dataset.
## Steps to reproduce the bug
```python
from datasets import load_dataset
datasets = load_dataset("nielsr/XFUN", "xfun.ja")
```
## Expected resu... | https://github.com/huggingface/datasets/issues/4099 | [
"Hi @andreybond, thanks for reporting.\r\n\r\nUnfortunately, I'm not able to able to reproduce your issue:\r\n```python\r\nIn [4]: from datasets import load_dataset\r\n ...: datasets = load_dataset(\"nielsr/XFUN\", \"xfun.ja\")\r\n\r\nIn [5]: datasets\r\nOut[5]: \r\nDatasetDict({\r\n train: Dataset({\r\n ... | null | 4,099 | false |
Proposing WikiSplit metric card | Pinging @lhoestq to ensure that my distinction between the dataset and the metric are clear :sweat_smile: | https://github.com/huggingface/datasets/pull/4098 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"A quick Github tip ;) To avoid running N times the CI, you can push all the changes at once: go to Files Changed tab, and on each suggestion there's a \"add to commit batch\" and then you can do one commit for all the suggestions you... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4098",
"html_url": "https://github.com/huggingface/datasets/pull/4098",
"diff_url": "https://github.com/huggingface/datasets/pull/4098.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4098.patch",
"merged_at": "2022-04-05T15:42... | 4,098 | true |
Updating FrugalScore metric card | removing duplicate paragraph | https://github.com/huggingface/datasets/pull/4097 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4097",
"html_url": "https://github.com/huggingface/datasets/pull/4097",
"diff_url": "https://github.com/huggingface/datasets/pull/4097.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4097.patch",
"merged_at": "2022-04-05T15:01... | 4,097 | true |
Add support for streaming Zarr stores for hosted datasets | **Is your feature request related to a problem? Please describe.**
Lots of geospatial data is stored in the Zarr format. This format works well for n-dimensional data and coordinates, and can have good compression. Unfortunately, HF datasets doesn't support streaming in data in Zarr format as far as I can tell. Zarr s... | https://github.com/huggingface/datasets/issues/4096 | [
"Hi @jacobbieker, thanks for your request and study of possible alternatives.\r\n\r\nWe are very interested in finding a way to make `datasets` useful to you.\r\n\r\nLooking at the Zarr docs, I saw that among its storage alternatives, there is the ZIP file format: https://zarr.readthedocs.io/en/stable/api/storage.h... | null | 4,096 | false |
fix typo in rename_column error message | I feel bad submitting such a tiny change as a PR but it confused me today ๐ | https://github.com/huggingface/datasets/pull/4095 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4095). All of your documentation changes will be reflected on that endpoint."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4095",
"html_url": "https://github.com/huggingface/datasets/pull/4095",
"diff_url": "https://github.com/huggingface/datasets/pull/4095.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4095.patch",
"merged_at": "2022-04-05T08:45... | 4,095 | true |
Helo Mayfrends | ## Adding a Dataset
- **Name:** *name of the dataset*
- **Description:** *short description of the dataset (or link to social media or blog post)*
- **Paper:** *link to the dataset paper if available*
- **Data:** *link to the Github repository or current dataset location*
- **Motivation:** *what are some good reas... | https://github.com/huggingface/datasets/issues/4094 | [] | null | 4,094 | false |
elena-soare/crawled-ecommerce: missing dataset | elena-soare/crawled-ecommerce
**Link:** *link to the dataset viewer page*
*short description of the issue*
Am I the one who added this dataset ? Yes-No
| https://github.com/huggingface/datasets/issues/4093 | [
"It's a bug! Thanks for reporting, I'm looking at it.",
"By the way, the error on our part is due to the huge size of every row (~90MB). The dataset viewer does not support such big dataset rows for the moment.\r\nAnyway, we're working to give a hint about this in the dataset viewer.",
"Fixed. See https://huggi... | null | 4,093 | false |
Fix dataset `amazon_us_reviews` metadata - 4/4/2022 | Fixes #4048 by running `dataset-cli test` to reprocess data and regenerate metadata. Additionally I've updated the README to include up-to-date counts for the subsets. | https://github.com/huggingface/datasets/pull/4092 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"cc: @albertvillanova just FYI"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4092",
"html_url": "https://github.com/huggingface/datasets/pull/4092",
"diff_url": "https://github.com/huggingface/datasets/pull/4092.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4092.patch",
"merged_at": "2022-04-08T12:29... | 4,092 | true |
Build a Dataset One Example at a Time Without Loading All Data Into Memory | **Is your feature request related to a problem? Please describe.**
I have a very large dataset stored on disk in a custom format. I have some custom code that reads one data example at a time and yields it in the form of a dictionary. I want to construct a `Dataset` with all examples, and then save it to disk. I la... | https://github.com/huggingface/datasets/issues/4091 | [
"Hi! Yes, the problem with `add_item` is that it keeps examples in memory, so you are left with these options:\r\n* writing a dataset loading script in which you iterate over `custom_example_dict_streamer` and yield the examples (in `_generate examples`)\r\n* storing the data in a JSON/CSV/Parquet/TXT file and usin... | null | 4,091 | false |
Avoid writing empty license files | This PR avoids the creation of empty `LICENSE` files. | https://github.com/huggingface/datasets/pull/4090 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4090",
"html_url": "https://github.com/huggingface/datasets/pull/4090",
"diff_url": "https://github.com/huggingface/datasets/pull/4090.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4090.patch",
"merged_at": "2022-04-07T12:40... | 4,090 | true |
Create metric card for Frugal Score | Proposing metric card for Frugal Score.
@albertvillanova or @lhoestq -- there are certain aspects that I'm not 100% sure on (such as how exactly the distillation between BertScore and FrugalScore is done) -- so if you find that something isn't clear, please let me know! | https://github.com/huggingface/datasets/pull/4089 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4089",
"html_url": "https://github.com/huggingface/datasets/pull/4089",
"diff_url": "https://github.com/huggingface/datasets/pull/4089.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4089.patch",
"merged_at": "2022-04-05T14:06... | 4,089 | true |
Remove unused legacy Beam utils | This PR removes unused legacy custom `WriteToParquet`, once official Apache Beam includes the patch since version 2.22.0:
- Patch PR: https://github.com/apache/beam/pull/11699
- Issue: https://issues.apache.org/jira/browse/BEAM-10022
In relation with:
- #204 | https://github.com/huggingface/datasets/pull/4088 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4088",
"html_url": "https://github.com/huggingface/datasets/pull/4088",
"diff_url": "https://github.com/huggingface/datasets/pull/4088.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4088.patch",
"merged_at": "2022-04-05T15:17... | 4,088 | true |
Fix BeamWriter output Parquet file | Since now, the `BeamWriter` saved a Parquet file with a simplified schema, where each field value was serialized to JSON. That resulted in Parquet files larger than Arrow files.
This PR:
- writes Parquet file preserving original schema and without serialization, thus avoiding serialization overhead and resulting in... | https://github.com/huggingface/datasets/pull/4087 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4087",
"html_url": "https://github.com/huggingface/datasets/pull/4087",
"diff_url": "https://github.com/huggingface/datasets/pull/4087.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4087.patch",
"merged_at": "2022-04-05T14:54... | 4,087 | true |
Dataset viewer issue for McGill-NLP/feedbackQA | ## Dataset viewer issue for '*McGill-NLP/feedbackQA*'
**Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/McGill-NLP/feedbackQA)*
*short description of the issue*
The dataset can be loaded correctly with `load_dataset` but the preview doesn't work. Error message:
```
Status code: 4... | https://github.com/huggingface/datasets/issues/4086 | [
"Hi @cslizc, thanks for reporting.\r\n\r\nI have just forced the refresh of the corresponding cache and the preview is working now.",
"thank you so much"
] | null | 4,086 | false |
datasets.set_progress_bar_enabled(False) not working in datasets v2 | ## Describe the bug
datasets.set_progress_bar_enabled(False) not working in datasets v2
## Steps to reproduce the bug
```python
datasets.set_progress_bar_enabled(False)
```
## Expected results
datasets not using any progress bar
## Actual results
AttributeError: module 'datasets' has no attribute 'se... | https://github.com/huggingface/datasets/issues/4085 | [
"Now, I can't find any reference to set_progress_bar_enabled in the code.\r\n\r\nI think it have been deleted",
"Hi @virilo,\r\n\r\nPlease note that since `datasets` version 2.0.0, we have aligned with `transformers` the management of the progress bar (among other things):\r\n- #3897\r\n\r\nNow, you should update... | null | 4,085 | false |
Errors in `Train with Datasets` Tensorflow code section on Huggingface.co | ## Describe the bug
Hi
### Error 1
Running the Tensforlow code on [Huggingface](https://huggingface.co/docs/datasets/use_dataset) gives a TypeError: __init__() got an unexpected keyword argument 'return_tensors'
### Error 2
`DataCollatorWithPadding` isn't imported
## Steps to reproduce the bug
```python
impo... | https://github.com/huggingface/datasets/issues/4084 | [
"Hi @blackhat-coder, thanks for reporting.\r\n\r\nPlease note that the `transformers` library updated their data collators API last year (version 4.10.0):\r\n- huggingface/transformers#13105\r\n\r\nnow requiring to pass `return_tensors` argument at Data Collator instantiation.\r\n\r\nAnd therefore, we also updated ... | null | 4,084 | false |
Add SacreBLEU Metric Card | null | https://github.com/huggingface/datasets/pull/4083 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4083",
"html_url": "https://github.com/huggingface/datasets/pull/4083",
"diff_url": "https://github.com/huggingface/datasets/pull/4083.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4083.patch",
"merged_at": "2022-04-12T20:38... | 4,083 | true |
Add chrF(++) Metric Card | null | https://github.com/huggingface/datasets/pull/4082 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4082",
"html_url": "https://github.com/huggingface/datasets/pull/4082",
"diff_url": "https://github.com/huggingface/datasets/pull/4082.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4082.patch",
"merged_at": "2022-04-12T20:38... | 4,082 | true |
Close parquet writer properly in `push_to_hub` | We donโt call writer.close(), which causes https://github.com/huggingface/datasets/issues/4077. It can happen that we upload the file before the writer is garbage collected and writes the footer.
I fixed this by explicitly closing the parquet writer.
Close https://github.com/huggingface/datasets/issues/4077. | https://github.com/huggingface/datasets/pull/4081 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@lhoestq / @albertvillanova / @mariosasko \r\nI am facing the same scenario. Let me explain the situation point. I have a glue ETL job\r\n\r\n1--> My files are in parquet format and stored in AWS s3.\r\n2--> I am iterating a loop fo... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4081",
"html_url": "https://github.com/huggingface/datasets/pull/4081",
"diff_url": "https://github.com/huggingface/datasets/pull/4081.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4081.patch",
"merged_at": "2022-04-01T16:16... | 4,081 | true |
NonMatchingChecksumError for downloading conll2012_ontonotesv5 dataset | ## Steps to reproduce the bug
```python
datasets.load_dataset("conll2012_ontonotesv5", "english_v12")
```
## Actual results
```
Downloading builder script: 32.2kB [00:00, 9.72MB/s]
Downloading metadata: 20.0kB [00:00, 10... | https://github.com/huggingface/datasets/issues/4080 | [
"Hi @richarddwang,\r\n\r\n\r\nIndeed, we have recently updated the loading script of that dataset (and fixed that bug as well):\r\n- #4002\r\n\r\nThat fix will be available in our next `datasets` library release. In the meantime, you can incorporate that fix by:\r\n- installing `datasets` from our GitHub repo:\r\n`... | null | 4,080 | false |
Increase max retries for GitHub datasets | As GitHub recurrently raises connectivity issues, this PR increases the number of max retries to request GitHub datasets, as previously done for GitHub metrics:
- #4063
Note that this is a temporary solution, while we decide when and how to load GitHub datasets from the Hub:
- #4059
Fix #2048
Related to:
- ... | https://github.com/huggingface/datasets/pull/4079 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4079",
"html_url": "https://github.com/huggingface/datasets/pull/4079",
"diff_url": "https://github.com/huggingface/datasets/pull/4079.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4079.patch",
"merged_at": "2022-04-01T15:27... | 4,079 | true |
Fix GithubMetricModuleFactory instantiation with None download_config | Recent PR:
- #4063
introduced a potential bug if `GithubMetricModuleFactory` is instantiated with None `download_config`.
This PR add instantiation tests and fix that potential issue.
CC: @lhoestq | https://github.com/huggingface/datasets/pull/4078 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4078",
"html_url": "https://github.com/huggingface/datasets/pull/4078",
"diff_url": "https://github.com/huggingface/datasets/pull/4078.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4078.patch",
"merged_at": "2022-04-01T14:39... | 4,078 | true |
ArrowInvalid: Parquet magic bytes not found in footer. Either the file is corrupted or this is not a parquet file. | ## Describe the bug
When uploading a relatively large image dataset of > 1GB, reloading doesn't work for me, even though pushing to the hub went just fine.
Basically, I do:
```
from datasets import load_dataset
dataset = load_dataset("imagefolder", data_files="path_to_my_files")
dataset.push_to_hub("dat... | https://github.com/huggingface/datasets/issues/4077 | [] | null | 4,077 | false |
Add ROUGE Metric Card | Add ROUGE metric card.
I've left the 'Values from popular papers' section empty for the time being because I don't know the summarization literature very well and am therefore not sure which paper(s) to pull from (note that the original rouge paper does not seem to present specific values, just correlations with hum... | https://github.com/huggingface/datasets/pull/4076 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4076",
"html_url": "https://github.com/huggingface/datasets/pull/4076",
"diff_url": "https://github.com/huggingface/datasets/pull/4076.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4076.patch",
"merged_at": "2022-04-12T20:37... | 4,076 | true |
Add CCAgT dataset | ## Adding a Dataset
- **Name:** CCAgT dataset: Images of Cervical Cells with AgNOR Stain Technique
- **Description:** The dataset contains 2540 images (1600x1200 where each pixel is 0.111ฮผmร0.111ฮผm) from three different slides, having at least one nucleus per image. These images are from fields belonging to a sample ... | https://github.com/huggingface/datasets/issues/4075 | [
"Awesome ! Let us know if you have questions or if we can help ;) I'm assigning you\r\n\r\nPS: if possible, please try to not use Google Drive links in your dataset script, since Google Drive has download quotas and is not always reliable.",
"HI, I was waiting to come out in the second version to do the implement... | null | 4,075 | false |
Error in google/xtreme_s dataset card | **Link:** https://huggingface.co/datasets/google/xtreme_s
Not a big deal but Hungarian is considered an Eastern European language, together with Serbian, Slovak, Slovenian (all correctly categorized; Slovenia is mostly to the West of Hungary, by the way).
| https://github.com/huggingface/datasets/issues/4074 | [
"Hi @wranai, thanks for reporting.\r\n\r\nPlease note that the information about language families and groups is taken form the original paper: [XTREME-S: Evaluating Cross-lingual Speech Representations](https://arxiv.org/abs/2203.10752).\r\n\r\nIf that information is wrong, feel free to contact the paper's authors... | null | 4,074 | false |
Create a metric card for Competition MATH | Proposing metric card for Competition MATH | https://github.com/huggingface/datasets/pull/4073 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4073",
"html_url": "https://github.com/huggingface/datasets/pull/4073",
"diff_url": "https://github.com/huggingface/datasets/pull/4073.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4073.patch",
"merged_at": "2022-04-01T18:57... | 4,073 | true |
Add installation instructions to image_process doc | This PR adds the installation instructions for the Image feature to the image process doc. | https://github.com/huggingface/datasets/pull/4072 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4072",
"html_url": "https://github.com/huggingface/datasets/pull/4072",
"diff_url": "https://github.com/huggingface/datasets/pull/4072.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4072.patch",
"merged_at": "2022-03-31T17:00... | 4,072 | true |
Loading issue for xuyeliu/notebookCDG dataset | ## Dataset viewer issue for '*xuyeliu/notebookCDG*'
**Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/xuyeliu/notebookCDG)*
*Couldn't load the xuyeliu/notebookCDG with provided scripts: *
```
from datasets import load_dataset
dataset = load_dataset("xuyeliu/notebookCDG/dataset_note... | https://github.com/huggingface/datasets/issues/4071 | [
"Hi @Jun-jie-Huang,\r\n\r\nAs the error message says, \".pkl\" data files are not supported.\r\n\r\nIf you would like to share your dataset on the Hub, you would need:\r\n- either to create a Python loading script, that loads the data in any format\r\n- or to transform your data files to one of the supported format... | null | 4,071 | false |
Create metric card for seqeval | Proposing metric card for seqeval. Not sure which values to report for Popular papers though. | https://github.com/huggingface/datasets/pull/4070 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4070",
"html_url": "https://github.com/huggingface/datasets/pull/4070",
"diff_url": "https://github.com/huggingface/datasets/pull/4070.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4070.patch",
"merged_at": "2022-04-01T18:57... | 4,070 | true |
Add support for metadata files to `imagefolder` | This PR adds support for metadata files to `imagefolder` to add an ability to specify image fields other than `image` and `label`, which are inferred from the directory structure in the loaded dataset.
To be parsed as an image metadata file, a file should be named `"info.csv"` and should have the following structure... | https://github.com/huggingface/datasets/pull/4069 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Love it !\r\n\r\n+1 to using JSON Lines rather than CSV. I've also seen image datasets for which JSON Lines was used.\r\n\r\nA `file_name` column sounds good as well, and it means we could reuse the same name for audio. And ok to che... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4069",
"html_url": "https://github.com/huggingface/datasets/pull/4069",
"diff_url": "https://github.com/huggingface/datasets/pull/4069.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4069.patch",
"merged_at": "2022-05-03T12:42... | 4,069 | true |
Improve out of bounds error message | In 1.18.4 with https://github.com/huggingface/datasets/pull/3719 we introduced an error message for users using `select` with out of bounds indices. The message ended up being confusing for some users because it mentioned negative indices, which is not the main use case.
I replaced it with a message that is very sim... | https://github.com/huggingface/datasets/pull/4068 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4068",
"html_url": "https://github.com/huggingface/datasets/pull/4068",
"diff_url": "https://github.com/huggingface/datasets/pull/4068.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4068.patch",
"merged_at": "2022-03-31T08:33... | 4,068 | true |
Update datasets task tags to align tags with models | **Requires https://github.com/huggingface/datasets/pull/4066 to be merged first**
Following https://github.com/huggingface/datasets/pull/4066 we need to update many dataset tags to use the new ones. This PR takes case of this and is quite big - feel free to review only certain tags if you don't want to spend too muc... | https://github.com/huggingface/datasets/pull/4067 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Looks good, but I think we are missing some scripts with outdated tags (RedCaps, MNIST, ...).",
"Just updated the tags of vision datasets :)\r\nWe can figure out one for image datasets without labels like PASS - not sure how to nam... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4067",
"html_url": "https://github.com/huggingface/datasets/pull/4067",
"diff_url": "https://github.com/huggingface/datasets/pull/4067.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4067.patch",
"merged_at": "2022-04-13T17:31... | 4,067 | true |
Tasks alignment with models | I updated our `tasks.json` file with the new task taxonomy that is aligned with models.
The rule that defines a task is the following:
**Two tasks are different if and only if the steps of their pipelines** are different, i.e. if they canโt reasonably be implemented using the same coherent code (level of granular... | https://github.com/huggingface/datasets/pull/4066 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Yay! This is exciting! Note that we would probably be able to generate this JSON directly from `huggingface/hub-docs`' `Types.ts` file (cc @osanseviero)",
"The following issue should make this much easier :smile: https://github.com... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4066",
"html_url": "https://github.com/huggingface/datasets/pull/4066",
"diff_url": "https://github.com/huggingface/datasets/pull/4066.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4066.patch",
"merged_at": "2022-04-08T12:20... | 4,066 | true |
Create metric card for METEOR | Proposing a metric card for METEOR | https://github.com/huggingface/datasets/pull/4065 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4065",
"html_url": "https://github.com/huggingface/datasets/pull/4065",
"diff_url": "https://github.com/huggingface/datasets/pull/4065.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4065.patch",
"merged_at": "2022-03-31T17:07... | 4,065 | true |
Contributing MedMCQA dataset | Adding MedMCQA dataset ( https://paperswithcode.com/dataset/medmcqa )
**Name**: MedMCQA
**Description**: MedMCQA is a large-scale, Multiple-Choice Question Answering (MCQA) dataset designed to address real-world medical entrance exam questions.
MedMCQA has more than 194k high-quality AIIMS & NEET PG entranc... | https://github.com/huggingface/datasets/pull/4064 | [
"@lhoestq Could you please take a look?\r\nThank you!!",
"Hi, thank you for the modifications and suggestions. Please check the changes.",
"Can you run `make style` to fix the code formatting please ?\r\n\r\nOh and was wrong with the dummy_data.zip file, it must actually be placed at `datasets/medmcqa/dummy/1.1... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4064",
"html_url": "https://github.com/huggingface/datasets/pull/4064",
"diff_url": "https://github.com/huggingface/datasets/pull/4064.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4064.patch",
"merged_at": "2022-05-06T08:42... | 4,064 | true |
Increase max retries for GitHub metrics | As GitHub recurrently raises connectivity issues, this PR increases the number of max retries to request GitHub metrics.
Related to:
- #3134
Also related to:
- #4059 | https://github.com/huggingface/datasets/pull/4063 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4063",
"html_url": "https://github.com/huggingface/datasets/pull/4063",
"diff_url": "https://github.com/huggingface/datasets/pull/4063.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4063.patch",
"merged_at": "2022-03-31T14:37... | 4,063 | true |
Loading mozilla-foundation/common_voice_7_0 dataset failed | ## Describe the bug
I wanted to load `mozilla-foundation/common_voice_7_0` dataset with `fi` language and `test` split from datasets on Colab/Kaggle notebook, but I am getting an error `JSONDecodeError: [Errno Expecting value] Not Found: 0` while loading it. The bug seems to affect other languages and splits too than ... | https://github.com/huggingface/datasets/issues/4062 | [
"Hi @aapot, thanks for reporting.\r\n\r\nWe are investigating the cause of this issue. We will keep you informed. ",
"When making HTTP request from code line:\r\n```\r\nresponse = requests.get(f\"{_API_URL}/bucket/dataset/{path}/{use_cdn}\", timeout=10.0).json()\r\n```\r\nit cannot be decoded to JSON because it r... | null | 4,062 | false |
Loading cnn_dailymail dataset failed | ## Describe the bug
I wanted to load cnn_dailymail dataset from huggingface datasets on jupyter lab, but I am getting an error ` NotADirectoryError:[Errno20] Not a directory ` while loading it.
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset('cnn_dailymail', '3.0.... | https://github.com/huggingface/datasets/issues/4061 | [
"Hi @Arij-Aladel, thanks for reporting.\r\n\r\nThis issue was already reported \r\n- #3784\r\n\r\nand its root cause is a change in the Google Drive service. See:\r\n- #3786 \r\n\r\nWe have already fixed it in our 2.0.0 release. See:\r\n- #3787 \r\n\r\nPlease, update your `datasets` version:\r\n```\r\npip install -... | null | 4,061 | false |
Deprecate canonical Multilingual Librispeech | Deprecate canonical Multilingual Librispeech in favor of [the community one](https://huggingface.co/datasets/facebook/multilingual_librispeech) which supports streaming.
However, there is a problem regarding new ASR template schema: since it's changed, I guess all community datasets that use this template do not wor... | https://github.com/huggingface/datasets/pull/4060 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Yes, as discussed in #4006 we should update facebook/multilingual_librispeech indeed before we do a release. @anton-l could you help taking care of updating facebook/multilingual_librispeech ? We need to update the task template\r\n`... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4060",
"html_url": "https://github.com/huggingface/datasets/pull/4060",
"diff_url": "https://github.com/huggingface/datasets/pull/4060.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4060.patch",
"merged_at": "2022-04-01T12:48... | 4,060 | true |
Load GitHub datasets from Hub | We have recurrently had connection errors when requesting GitHub because sometimes the site is not available.
This PR requests the Hub instead, once all GitHub datasets are mirrored on the Hub.
Fix #2048
Related to:
- #4051
- #3210
- #2787
- #2075
- #2036 | https://github.com/huggingface/datasets/pull/4059 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Currently the github datasets versioning is synced with the `datasets` lib versioning: when you load a github dataset using `datasets==x.y.z`, then the version of the dataset will be the one at the git tag `x.y.z`. This is for reprod... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4059",
"html_url": "https://github.com/huggingface/datasets/pull/4059",
"diff_url": "https://github.com/huggingface/datasets/pull/4059.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4059.patch",
"merged_at": "2022-09-16T12:40... | 4,059 | true |
Updated annotations for nli_tr dataset | This PR adds annotation tags for `nli_tr` dataset so that the dataset can be searchable wrt. relevant query parameters.
The annotations in this PR are based on the existing annotations of `snli` and `multi_nli` datasets as `nli_tr` is a machine-generated extension of those datasets.
This PR is intended only for u... | https://github.com/huggingface/datasets/pull/4058 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thank you so much @[lhoestq](https://github.com/lhoestq) for the time you take to your review the PR!"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4058",
"html_url": "https://github.com/huggingface/datasets/pull/4058",
"diff_url": "https://github.com/huggingface/datasets/pull/4058.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4058.patch",
"merged_at": "2022-04-12T10:37... | 4,058 | true |
`load_dataset` consumes too much memory for audio + tar archives |
## Description
`load_dataset` consumes more and more memory until it's killed, even though it's made with a generator. I'm adding a loading script for a new dataset, made up of ~15s audio coming from a tar file. Tried setting `DEFAULT_WRITER_BATCH_SIZE = 1` as per the discussion in #741 but the problem persists.
... | https://github.com/huggingface/datasets/issues/4057 | [
"Hi ! Could it be because you need to free the memory used by `tarfile` by emptying the tar `members` by any chance ?\r\n```python\r\n yield key, {\"audio\": {\"path\": audio_name, \"bytes\": audio_file_obj.read()}}\r\n audio_tarfile.members = [] # free memory\r\n key += 1\r\n```\r\n\r\nand th... | null | 4,057 | false |
Unexpected behavior of _TempDirWithCustomCleanup | ## Describe the bug
This is not 100% a bug in `datasets`, but behavior that surprised me and I think this could be made more robust on the `datasets`side.
When using `datasets.disable_caching()`, cache files are written to a temporary directory. This directory should be based on the environment variable TMPDIR. I ... | https://github.com/huggingface/datasets/issues/4056 | [
"Hi ! Would setting TMPDIR at the beginning of your python script/session work ? I mean, even before importing transformers, datasets, etc. and using them ? I think this would be the most robust solution given any library that uses `tempfile`. I don't think we aim to support environment variables to be changed at r... | null | 4,056 | false |
[DO NOT MERGE] Test doc-builder | This is a test PR to ensure the changes in https://github.com/huggingface/doc-builder/pull/164 don't break anything in `datasets` | https://github.com/huggingface/datasets/pull/4055 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Docs built successfully, so closing this."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4055",
"html_url": "https://github.com/huggingface/datasets/pull/4055",
"diff_url": "https://github.com/huggingface/datasets/pull/4055.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4055.patch",
"merged_at": null
} | 4,055 | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.