title
stringlengths
1
290
body
stringlengths
0
228k
html_url
stringlengths
46
51
comments
list
pull_request
dict
number
int64
1
5.59k
is_pull_request
bool
2 classes
Issues with concatenating datasets
## Describe the bug It is impossible to concatenate datasets if a feature is sequence of dict in one dataset and a dict of sequence in another. But based on the document, it should be automatically converted. > A [datasets.Sequence](https://huggingface.co/docs/datasets/v2.3.2/en/package_reference/main_classes#datas...
https://github.com/huggingface/datasets/issues/4666
[ "Hi! I agree we should improve the features equality checks to account for this particular case. However, your code fails due to `answer_start` having the dtype `int64` instead of `int32` after loading from JSON (it's not possible to embed type precision info into a JSON file; `save_to_disk` does that for arrow fil...
null
4,666
false
Unable to create dataset having Python dataset script only
## Describe the bug Hi there, I'm trying to add the following dataset to Huggingface datasets: https://huggingface.co/datasets/Heriot-WattUniversity/dialog-babi/blob/ I'm trying to do so using the CLI commands but seems that this command generates the wrong `dataset_info.json` file (you can find it in the repo a...
https://github.com/huggingface/datasets/issues/4665
[ "Hi @aleSuglia, thanks for reporting.\r\n\r\nWe are having a look at it. \r\n\r\nWe transfer this issue to the Community tab of the corresponding Hub dataset: https://huggingface.co/datasets/Heriot-WattUniversity/dialog-babi/discussions" ]
null
4,665
false
Add stanford dog dataset
This PR is for adding dataset, related to issue #4504. We are adding Stanford dog breed dataset. It is a multi class image classification dataset. Details can be found here - http://vision.stanford.edu/aditya86/ImageNetDogs/ Tests on dummy data is failing currently, which I am looking into.
https://github.com/huggingface/datasets/pull/4664
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi @khushmeeet, thanks for your contribution.\r\n\r\nBut wouldn't it be better to add this dataset to the Hub? \r\n- https://huggingface.co/docs/datasets/share\r\n- https://huggingface.co/docs/datasets/dataset_script", "Hi @albertv...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4664", "html_url": "https://github.com/huggingface/datasets/pull/4664", "diff_url": "https://github.com/huggingface/datasets/pull/4664.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4664.patch", "merged_at": null }
4,664
true
Add text decorators
This PR adds some decoration to text about different modalities to make it more obvious separate guides exist for audio, vision, and text. The goal is to make it easier for users to discover these guides! ![underline](https://user-images.githubusercontent.com/59462357/178044392-9596693e-9a4a-479a-a282-f1edbd90be1a.p...
https://github.com/huggingface/datasets/pull/4663
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4663", "html_url": "https://github.com/huggingface/datasets/pull/4663", "diff_url": "https://github.com/huggingface/datasets/pull/4663.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4663.patch", "merged_at": "2022-07-18T18:20...
4,663
true
Fix: conll2003 - fix empty example
As reported in https://huggingface.co/datasets/conll2003/discussions/2#62c45a14f93fc97e8260532f, there was an extra empty example at the end of the dataset
https://github.com/huggingface/datasets/pull/4662
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4662", "html_url": "https://github.com/huggingface/datasets/pull/4662", "diff_url": "https://github.com/huggingface/datasets/pull/4662.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4662.patch", "merged_at": "2022-07-08T14:02...
4,662
true
Concurrency bug when using same cache among several jobs
## Describe the bug I used to see this bug with an older version of the datasets. It seems to persist. This is my concrete scenario: I launch several evaluation jobs on a cluster in which I share the file system and I share the cache directory used by huggingface libraries. The evaluation jobs read the same *.csv ...
https://github.com/huggingface/datasets/issues/4661
[ "I can confirm that if I run one job first that processes the dataset, then I can run any jobs in parallel with no problem (no write-concurrency anymore...). ", "Hi! That's weird. It seems like the error points to the `mkstemp` function, but the official docs state the following:\r\n```\r\nThere are no race condi...
null
4,661
false
Fix _resolve_single_pattern_locally on Windows with multiple drives
Currently, when `_resolve_single_pattern_locally` is called from a different drive than the one in `pattern`, it raises an exception: ``` _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ C:\hostedtoolcache\windows\Python\3.6.8\x64\lib\site-packages\datasets\io\parquet.py:35: in __init_...
https://github.com/huggingface/datasets/pull/4660
[ "_The documentation is not available anymore as the PR was closed or merged._", "Good catch ! Sorry I forgot (again) about windows paths when writing this x)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4660", "html_url": "https://github.com/huggingface/datasets/pull/4660", "diff_url": "https://github.com/huggingface/datasets/pull/4660.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4660.patch", "merged_at": "2022-07-07T16:52...
4,660
true
Transfer CI to GitHub Actions
This PR transfers CI from CircleCI to GitHub Actions. The implementation in GitHub Actions tries to be as faithful as possible to the implementation in CircleCI and get the same output results (exceptions below). **IMPORTANT NOTE**: The fast-fail policy (described below) is not finally implemented, so that: - we c...
https://github.com/huggingface/datasets/pull/4659
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thanks a lot @albertvillanova ! I hope we're finally done with flakiness on windows ^^\r\n\r\nAlso thanks for paying extra attention to billing and avoiding running unnecessary jobs. Though for certain aspects (see my comments), I th...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4659", "html_url": "https://github.com/huggingface/datasets/pull/4659", "diff_url": "https://github.com/huggingface/datasets/pull/4659.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4659.patch", "merged_at": "2022-07-12T11:18...
4,659
true
Transfer CI tests to GitHub Actions
Let's try CI tests using GitHub Actions to see if they are more stable than on CircleCI.
https://github.com/huggingface/datasets/issues/4658
[]
null
4,658
false
Add SQuAD2.0 Dataset
## Adding a Dataset - **Name:** *SQuAD2.0* - **Description:** *Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading ...
https://github.com/huggingface/datasets/issues/4657
[ "Hey, It's already present [here](https://huggingface.co/datasets/squad_v2) ", "Hi! This dataset is indeed already available on the Hub. Closing." ]
null
4,657
false
Add Amazon-QA Dataset
## Adding a Dataset - **Name:** *Amazon-QA* - **Description:** *The dataset is .jsonl format, where each line in the file is a json string that corresponds to a question, existing answers to the question and the extracted review snippets (relevant to the question).* - **Paper:** *https://github.com/amazonqa/amazonqa...
https://github.com/huggingface/datasets/issues/4656
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/Amazon-QA)." ]
null
4,656
false
Simple Wikipedia
## Adding a Dataset - **Name:** *Simple Wikipedia* - **Description:** *Two different versions of the data set now exist. Both were generated by aligning Simple English Wikipedia and English Wikipedia. A complete description of the extraction process can be found in "Simple English Wikipedia: A New Simplification Task...
https://github.com/huggingface/datasets/issues/4655
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/simple-wiki)." ]
null
4,655
false
Add Quora Question Triplets Dataset
## Adding a Dataset - **Name:** *Quora Question Triplets* - **Description:** *This dataset consists of over 400,000 lines of potential question duplicate pairs. Each line contains IDs for each question in the pair, the full text for each question, and a binary value that indicates whether the line truly contains a du...
https://github.com/huggingface/datasets/issues/4654
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/QQP_triplets)." ]
null
4,654
false
Add Altlex dataset
## Adding a Dataset - **Name:** *Altlex* - **Description:** *Git repository for software associated with the 2016 ACL paper "Identifying Causal Relations Using Parallel Wikipedia Articles.”* - **Paper:** *https://aclanthology.org/P16-1135.pdf* - **Data:** *https://huggingface.co/datasets/sentence-transformers/embed...
https://github.com/huggingface/datasets/issues/4653
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/altlex)." ]
null
4,653
false
Add Sentence Compression Dataset
## Adding a Dataset - **Name:** *Sentence Compression* - **Description:** *Large corpus of uncompressed and compressed sentences from news articles.* - **Paper:** *https://www.aclweb.org/anthology/D13-1155/* - **Data:** *https://github.com/google-research-datasets/sentence-compression/tree/master/data* - **Motivat...
https://github.com/huggingface/datasets/issues/4652
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/sentence-compression)." ]
null
4,652
false
Add Flickr 30k Dataset
## Adding a Dataset - **Name:** *Flickr 30k* - **Description:** *To produce the denotation graph, we have created an image caption corpus consisting of 158,915 crowd-sourced captions describing 31,783 images. This is an extension of our previous Flickr 8k Dataset. The new images and captions focus on people involved ...
https://github.com/huggingface/datasets/issues/4651
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/flickr30k-captions)." ]
null
4,651
false
Add SPECTER dataset
## Adding a Dataset - **Name:** *SPECTER* - **Description:** *SPECTER: Document-level Representation Learning using Citation-informed Transformers* - **Paper:** *https://doi.org/10.18653/v1/2020.acl-main.207* - **Data:** *https://huggingface.co/datasets/sentence-transformers/embedding-training-data/resolve/main/spe...
https://github.com/huggingface/datasets/issues/4650
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/SPECTER)" ]
null
4,650
false
Add PAQ dataset
## Adding a Dataset - **Name:** *PAQ* - **Description:** *This repository contains code and models to support the research paper PAQ: 65 Million Probably-Asked Questions and What You Can Do With Them* - **Paper:** *https://arxiv.org/abs/2102.07033* - **Data:** *https://huggingface.co/datasets/sentence-transformers/...
https://github.com/huggingface/datasets/issues/4649
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/PAQ_pairs)" ]
null
4,649
false
Add WikiAnswers dataset
## Adding a Dataset - **Name:** *WikiAnswers* - **Description:** *The WikiAnswers corpus contains clusters of questions tagged by WikiAnswers users as paraphrases. Each cluster optionally contains an answer provided by WikiAnswers users.* - **Paper:** *https://dl.acm.org/doi/10.1145/2623330.2623677* - **Data:** *ht...
https://github.com/huggingface/datasets/issues/4648
[ "uploaded dataset [here](https://huggingface.co/datasets/embedding-data/WikiAnswers)" ]
null
4,648
false
Add Reddit dataset
## Adding a Dataset - **Name:** *Reddit comments (2015-2018)* - **Description:** *Reddit is an American social news aggregation website, where users can post links, and take part in discussions on these posts. These threaded discussions provide a large corpus, which is converted into a conversational dataset using th...
https://github.com/huggingface/datasets/issues/4647
[]
null
4,647
false
Set HF_SCRIPTS_VERSION to main
After renaming "master" to "main", the CI fails with ``` AssertionError: 'https://raw.githubusercontent.com/huggingface/datasets/main/datasets/_dummy/_dummy.py' not found in "Couldn't find a dataset script at /home/circleci/datasets/_dummy/_dummy.py or any data file in the same directory. Couldn't find '_dummy' on th...
https://github.com/huggingface/datasets/pull/4645
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4645", "html_url": "https://github.com/huggingface/datasets/pull/4645", "diff_url": "https://github.com/huggingface/datasets/pull/4645.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4645.patch", "merged_at": "2022-07-06T15:45...
4,645
true
[Minor fix] Typo correction
recieve -> receive
https://github.com/huggingface/datasets/pull/4644
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4644", "html_url": "https://github.com/huggingface/datasets/pull/4644", "diff_url": "https://github.com/huggingface/datasets/pull/4644.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4644.patch", "merged_at": "2022-07-06T15:45...
4,644
true
Rename master to main
This PR renames mentions of "master" by "main" in the code base for several cases: - set the default dataset script version to "main" if the local installation of `datasets` is a dev installation - update URLs to this github repository to use "main" - update the DVC benchmark - update the github workflows - update...
https://github.com/huggingface/datasets/pull/4643
[ "_The documentation is not available anymore as the PR was closed or merged._", "All the mentions I found on google were simple URLs that will be redirected, so it's fine. I also checked the spaces and we should be good:\r\n- dalle-mini used to install the master branch but [it's no longer the case](https://huggi...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4643", "html_url": "https://github.com/huggingface/datasets/pull/4643", "diff_url": "https://github.com/huggingface/datasets/pull/4643.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4643.patch", "merged_at": "2022-07-06T15:25...
4,643
true
Streaming issue for ccdv/pubmed-summarization
### Link https://huggingface.co/datasets/ccdv/pubmed-summarization ### Description This was reported by a [user of AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/7). It seems like streaming doesn't work due to the way the dataset loading script is defined? ``` Status c...
https://github.com/huggingface/datasets/issues/4642
[ "Thanks for reporting @lewtun.\r\n\r\nI confirm there is an issue with streaming: it does not stream locally. ", "Oh, after investigation, the source of the issue is in the Hub dataset loading script.\r\n\r\nI'm opening a PR on the Hub dataset.", "I've opened a PR on their Hub dataset to support streaming: http...
null
4,642
false
Dataset Viewer issue for kmfoda/booksum
### Link https://huggingface.co/datasets/kmfoda/booksum ### Description A [user of AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/9) discovered this dataset cannot be streamed due to: ``` Status code: 400 Exception: ClientResponseError Message: 401, messa...
https://github.com/huggingface/datasets/issues/4641
[ "Thanks for reporting, @lewtun.\r\n\r\nIt works locally in streaming mode:\r\n```\r\n{'bid': 27681,\r\n 'is_aggregate': True,\r\n 'source': 'cliffnotes',\r\n 'chapter_path': 'all_chapterized_books/27681-chapters/chapters_1_to_2.txt',\r\n 'summary_path': 'finished_summaries/cliffnotes/The Last of the Mohicans/sectio...
null
4,641
false
Support all split in streaming mode
Fix #4637.
https://github.com/huggingface/datasets/pull/4640
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4640). All of your documentation changes will be reflected on that endpoint." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4640", "html_url": "https://github.com/huggingface/datasets/pull/4640", "diff_url": "https://github.com/huggingface/datasets/pull/4640.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4640.patch", "merged_at": null }
4,640
true
Add HaGRID -- HAnd Gesture Recognition Image Dataset
## Adding a Dataset - **Name:** HaGRID -- HAnd Gesture Recognition Image Dataset - **Description:** We introduce a large image dataset HaGRID (HAnd Gesture Recognition Image Dataset) for hand gesture recognition (HGR) systems. You can use it for image classification or image detection tasks. Proposed dataset allows t...
https://github.com/huggingface/datasets/issues/4639
[]
null
4,639
false
The speechocean762 dataset
[speechocean762](https://www.openslr.org/101/) is a non-native English corpus for pronunciation scoring tasks. It is free for both commercial and non-commercial use. I believe it will be easier to use if it could be available on Hugging Face.
https://github.com/huggingface/datasets/pull/4638
[ "CircleCL reported two errors, but I didn't find the reason. The error message:\r\n```\r\n_________________ ERROR collecting tests/test_dataset_cards.py _________________\r\ntests/test_dataset_cards.py:53: in <module>\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\ntests/test_d...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4638", "html_url": "https://github.com/huggingface/datasets/pull/4638", "diff_url": "https://github.com/huggingface/datasets/pull/4638.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4638.patch", "merged_at": null }
4,638
true
The "all" split breaks streaming
## Describe the bug Not sure if this is a bug or just the way streaming works, but setting `streaming=True` did not work when setting `split="all"` ## Steps to reproduce the bug The following works: ```python ds = load_dataset('super_glue', 'wsc.fixed', split='all') ``` The following throws `ValueError: Bad ...
https://github.com/huggingface/datasets/issues/4637
[ "Thanks for reporting @cakiki.\r\n\r\nYes, this is a bug. We are investigating it.", "@albertvillanova Nice! Let me know if it's something I can fix my self; would love to contribtue!", "@cakiki I was working on this but if you would like to contribute, go ahead. I will close my PR. ;)\r\n\r\nFor the moment I j...
null
4,637
false
Add info in docs about behavior of download_config.num_proc
**Is your feature request related to a problem? Please describe.** I went to override `download_config.num_proc` and was confused about what was happening under the hood. It would be nice to have the behavior documented a bit better so folks know what's happening when they use it. **Describe the solution you'd li...
https://github.com/huggingface/datasets/issues/4636
[]
null
4,636
false
Dataset Viewer issue for vadis/sv-ident
### Link https://huggingface.co/datasets/vadis/sv-ident/viewer/default/validation ### Description Error message when loading validation split in the viewer: ``` Status code: 400 Exception: Status400Error Message: The split cache is empty. ``` ### Owner _No response_
https://github.com/huggingface/datasets/issues/4635
[ "Thanks for reporting, @e-tornike \r\n\r\nSome context:\r\n- #4527 \r\n\r\nThe dataset loads locally in streaming mode:\r\n```python\r\nIn [2]: from datasets import load_dataset; ds = load_dataset(\"vadis/sv-ident\", split=\"validation\", streaming=True); item = next(iter(ds)); item\r\nUsing custom data configurati...
null
4,635
false
Can't load the Hausa audio dataset
common_voice_train = load_dataset("common_voice", "ha", split="train+validation")
https://github.com/huggingface/datasets/issues/4634
[ "Could you provide the error details. It is difficult to debug otherwise. Also try other config. `ha` is not a valid." ]
null
4,634
false
[data_files] Only match separated split names
As reported in https://github.com/huggingface/datasets/issues/4477, the current pattern matching to infer which file goes into which split is too permissive. For example a file "contest.py" would be considered part of a test split (it contains "test") and "seqeval.py" as well (it contains "eval"). In this PR I made ...
https://github.com/huggingface/datasets/pull/4633
[ "_The documentation is not available anymore as the PR was closed or merged._", "I ran a script to find affected datasets (just did it on non-private non-gated). Adding \"testing\" and \"evaluation\" fixes all of of them except one:\r\n- projecte-aina/cat_manynames:\thuman_annotated_testset.tsv\r\n\r\nLet me open...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4633", "html_url": "https://github.com/huggingface/datasets/pull/4633", "diff_url": "https://github.com/huggingface/datasets/pull/4633.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4633.patch", "merged_at": "2022-07-18T13:07...
4,633
true
'sort' method sorts one column only
The 'sort' method changes the order of one column only (the one defined by the argument 'column'), thus creating a mismatch between a sample fields. I would expect it to change the order of the samples as a whole, based on the 'column' order.
https://github.com/huggingface/datasets/issues/4632
[ "Hi ! `ds.sort()` does sort the full dataset, not just one column:\r\n```python\r\nfrom datasets import *\r\n\r\nds = Dataset.from_dict({\"foo\": [3, 2, 1], \"bar\": [\"c\", \"b\", \"a\"]})\r\nprint(d.sort(\"foo\").to_pandas()\r\n# foo bar\r\n# 0 1 a\r\n# 1 2 b\r\n# 2 3 c\r\n```\r\n\r\nWhat made y...
null
4,632
false
Update WinoBias README
I'm adding some information about Winobias that I got from the paper :smile: I think this makes it a bit clearer!
https://github.com/huggingface/datasets/pull/4631
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4631", "html_url": "https://github.com/huggingface/datasets/pull/4631", "diff_url": "https://github.com/huggingface/datasets/pull/4631.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4631.patch", "merged_at": "2022-07-07T13:11...
4,631
true
fix(dataset_wrappers): Fixes access to fsspec.asyn in torch_iterable_dataset.py.
Fix #4612. Apparently, newest `fsspec` versions do not allow access to attribute-based modules if they are not imported, such as `fsspec.async`. Thus, @mariosasko suggested to add the missing part to the module import to allow for its access.
https://github.com/huggingface/datasets/pull/4630
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4630", "html_url": "https://github.com/huggingface/datasets/pull/4630", "diff_url": "https://github.com/huggingface/datasets/pull/4630.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4630.patch", "merged_at": "2022-07-05T15:08...
4,630
true
Rename repo default branch to main
Rename repository default branch to `main` (instead of current `master`). Once renamed, users will have to manually update their local repos: - [ ] Upstream: ``` git branch -m master main git fetch upstream main git branch -u upstream/main main git remote set-head upstream -a ``` - [ ] Origin...
https://github.com/huggingface/datasets/issues/4629
[]
null
4,629
false
Fix time type `_arrow_to_datasets_dtype` conversion
Fix #4620 The issue stems from the fact that `pa.array([time_data]).type` returns `DataType(time64[unit])`, which doesn't expose the `unit` attribute, instead of `Time64Type(time64[unit])`. I believe this is a bug in PyArrow. Luckily, the both types have the same `str()`, so in this PR I call `pa.type_for_alias(str(...
https://github.com/huggingface/datasets/pull/4628
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4628", "html_url": "https://github.com/huggingface/datasets/pull/4628", "diff_url": "https://github.com/huggingface/datasets/pull/4628.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4628.patch", "merged_at": "2022-07-07T13:57...
4,628
true
fixed duplicate calculation of spearmanr function in metrics wrapper.
During _compute, the scipy.stats spearmanr function was called twice, redundantly, once for calculating the score and once for calculating the p-value, under the conditional branch where return_pvalue=True. I adjusted the _compute function to execute the spearmanr function once, store the results tuple in a temporary v...
https://github.com/huggingface/datasets/pull/4627
[ "Great, can open a PR in `evaluate` as well to optimize this.\r\n\r\nRelatedly, I wanted to add a new metric, Kendall Tau (https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.kendalltau.html). If I were to open a PR with the wrapper, description, citation, docstrings, readme, etc. would it make more se...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4627", "html_url": "https://github.com/huggingface/datasets/pull/4627", "diff_url": "https://github.com/huggingface/datasets/pull/4627.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4627.patch", "merged_at": "2022-07-07T12:41...
4,627
true
Add non-commercial licensing info for datasets for which we removed tags
We removed several YAML tags saying that certain datasets can't be used for commercial purposes: https://github.com/huggingface/datasets/pull/4613#discussion_r911919753 Reason for this is that we only allow tags that are part of our [supported list of licenses](https://github.com/huggingface/datasets/blob/84fc3ad73c...
https://github.com/huggingface/datasets/issues/4626
[ "yep plus `license_details` also makes sense for this IMO" ]
null
4,626
false
Unpack `dl_manager.iter_files` to allow parallization
Iterate over data files outside `dl_manager.iter_files` to allow parallelization in streaming mode. (The issue reported [here](https://discuss.huggingface.co/t/dataset-only-have-n-shard-1-when-has-multiple-shards-in-repo/19887)) PS: Another option would be to override `FilesIterable.__getitem__` to make it indexa...
https://github.com/huggingface/datasets/pull/4625
[ "_The documentation is not available anymore as the PR was closed or merged._", "Cool thanks ! Yup it sounds like the right solution.\r\n\r\nIt looks like `_generate_tables` needs to be updated as well to fix the CI" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4625", "html_url": "https://github.com/huggingface/datasets/pull/4625", "diff_url": "https://github.com/huggingface/datasets/pull/4625.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4625.patch", "merged_at": "2022-07-05T11:00...
4,625
true
Remove all paperswithcode_id: null
On the Hub there is a validation error on the `paperswithcode_id` tag when the value is `null`: <img width="686" alt="image" src="https://user-images.githubusercontent.com/42851186/177151825-93d341c5-25bd-41ab-96c2-c0b516d51c68.png"> We've been using `null` to specify that we checked on pwc but the dataset doesn'...
https://github.com/huggingface/datasets/pull/4624
[ "_The documentation is not available anymore as the PR was closed or merged._", "> We've been using `null` to specify that we checked on pwc but the dataset doesn't exist there.\r\n\r\n@lhoestq maybe it's better to accept it on the Hub side then? Let me know if you want us to do it Hub-side", "Yup it's maybe be...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4624", "html_url": "https://github.com/huggingface/datasets/pull/4624", "diff_url": "https://github.com/huggingface/datasets/pull/4624.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4624.patch", "merged_at": null }
4,624
true
Loading MNIST as Pytorch Dataset
## Describe the bug Conversion of MNIST dataset to pytorch fails with bug ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("mnist", split="train") dataset.set_format('torch') dataset[0] print() ``` ## Expected results Expect to see torch tensors image and l...
https://github.com/huggingface/datasets/issues/4623
[ "Hi ! We haven't implemented the conversion from images data to PyTorch tensors yet I think\r\n\r\ncc @mariosasko ", "So I understand:\r\n\r\nset_format() does not properly do the conversion to pytorch tensors from PIL images.\r\n\r\nSo that someone who stumbles on this can use the package:\r\n\r\n```python\r\nda...
null
4,623
false
Fix ImageFolder with parameters drop_metadata=True and drop_labels=False (when metadata.jsonl is present)
Will fix #4621 ImageFolder raises `KeyError: 'label'` with params `drop_metadata=True` and `drop_labels=False` (if there is at least one metadata.jsonl file a data directory). This happens because metadata files are collected inside `analyze()` function regardless of `drop_metadata` value. And then the following co...
https://github.com/huggingface/datasets/pull/4622
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq @mariosasko pls take a look at https://github.com/huggingface/datasets/pull/4622/commits/769e4c046a5bd5e3a4dbd09cfad1f4cf60677869. I modified `_generate_examples()` according to the same logic too: removed checking if `metad...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4622", "html_url": "https://github.com/huggingface/datasets/pull/4622", "diff_url": "https://github.com/huggingface/datasets/pull/4622.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4622.patch", "merged_at": "2022-07-15T14:24...
4,622
true
ImageFolder raises an error with parameters drop_metadata=True and drop_labels=False when metadata.jsonl is present
## Describe the bug If you pass `drop_metadata=True` and `drop_labels=False` when a `data_dir` contains at least one `matadata.jsonl` file, you will get a KeyError. This is probably not a very useful case but we shouldn't get an error anyway. Asking users to move metadata files manually outside `data_dir` or pass fe...
https://github.com/huggingface/datasets/issues/4621
[]
null
4,621
false
Data type is not recognized when using datetime.time
## Describe the bug Creating a dataset from a pandas dataframe with `datetime.time` format generates an error. ## Steps to reproduce the bug ```python import pandas as pd from datetime import time from datasets import Dataset df = pd.DataFrame({"feature_name": [time(1, 1, 1)]}) dataset = Dataset.from_pandas...
https://github.com/huggingface/datasets/issues/4620
[ "cc @mariosasko ", "Hi, thanks for reporting! I'm investigating the issue." ]
null
4,620
false
np arrays get turned into native lists
## Describe the bug When attaching an `np.array` field, it seems that it automatically gets turned into a list (see below). Why is this happening? Could it lose precision? Is there a way to make sure this doesn't happen? ## Steps to reproduce the bug ```python >>> import datasets, numpy as np >>> dataset = datas...
https://github.com/huggingface/datasets/issues/4619
[ "If you add the line `dataset2.set_format('np')` before calling `dataset2[0]['tmp']` it should return `np.ndarray`.\r\nI believe internally it will not store it as a list, it is only returning a list when you index it.\r\n\r\n```\r\nIn [1]: import datasets, numpy as np\r\nIn [2]: dataset = datasets.load_dataset(\"g...
null
4,619
false
contribute data loading for object detection datasets with yolo data format
**Is your feature request related to a problem? Please describe.** At the moment, HF datasets loads [image classification datasets](https://huggingface.co/docs/datasets/image_process) out-of-the-box. There could be a data loader for loading standard object detection datasets ([original discussion here](https://hugging...
https://github.com/huggingface/datasets/issues/4618
[ "Hi! The `imagefolder` script is already quite complex, so a standalone script sounds better. Also, I suggest we create an org on the Hub (e.g. `hf-loaders`) and store such scripts there for easier maintenance rather than having them as packaged modules (IMO only very generic loaders should be packaged). WDYT @lhoe...
null
4,618
false
Fix `embed_storage` on features inside lists/sequences
Add a dedicated function for embed_storage to always preserve the embedded/casted arrays (and to have more control over `embed_storage` in general). Fix #4591 ~~(Waiting for #4608 to be merged to mark this PR as ready for review - required for fixing `xgetsize` in private repos)~~ Done!
https://github.com/huggingface/datasets/pull/4615
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4615", "html_url": "https://github.com/huggingface/datasets/pull/4615", "diff_url": "https://github.com/huggingface/datasets/pull/4615.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4615.patch", "merged_at": "2022-07-08T12:01...
4,615
true
Ensure ConcatenationTable.cast uses target_schema metadata
Currently, `ConcatenationTable.cast` does not use target_schema metadata when casting subtables. This causes an issue when using cast_column and the underlying table is a ConcatenationTable. Code example of where issue arrises: ``` from datasets import Dataset, Image column1 = [0, 1] image_paths = ['/images/im...
https://github.com/huggingface/datasets/pull/4614
[ "Hi @lhoestq, Thanks for the detailed comment. I've tested the suggested approach and can confirm it works for the testcase outlined above! The PR is updated with the changes.", "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4614", "html_url": "https://github.com/huggingface/datasets/pull/4614", "diff_url": "https://github.com/huggingface/datasets/pull/4614.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4614.patch", "merged_at": "2022-07-19T13:36...
4,614
true
Align/fix license metadata info
fix bad "other-*" licenses and add the corresponding "license_details" when relevant
https://github.com/huggingface/datasets/pull/4613
[ "_The documentation is not available anymore as the PR was closed or merged._", "Thank you thank you! Let's merge and pray? 😱 ", "I just need to add `license_details` to the validator and yup we can merge" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4613", "html_url": "https://github.com/huggingface/datasets/pull/4613", "diff_url": "https://github.com/huggingface/datasets/pull/4613.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4613.patch", "merged_at": "2022-07-01T12:42...
4,613
true
Release 2.3.0 broke custom iterable datasets
## Describe the bug Trying to iterate examples from custom iterable dataset fails to bug introduced in `torch_iterable_dataset.py` since the release of 2.3.0. ## Steps to reproduce the bug ```python next(iter(custom_iterable_dataset)) ``` ## Expected results `next(iter(custom_iterable_dataset))` should retu...
https://github.com/huggingface/datasets/issues/4612
[ "Apparently, `fsspec` does not allow access to attribute-based modules anymore, such as `fsspec.async`.\r\n\r\nHowever, this is a fairly simple fix:\r\n- Change the import to: `from fsspec import asyn`;\r\n- Change line 18 to: `asyn.iothread[0] = None`;\r\n- Change line 19 to `asyn.loop[0] = None`.", "Hi! I think...
null
4,612
false
Preserve member order by MockDownloadManager.iter_archive
Currently, `MockDownloadManager.iter_archive` yields paths to archive members in an order given by `path.rglob("*")`, which migh not be the same order as in the original archive. See issue in: - https://github.com/huggingface/datasets/pull/4579#issuecomment-1172135027 This PR fixes the order of the members yield...
https://github.com/huggingface/datasets/pull/4611
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4611", "html_url": "https://github.com/huggingface/datasets/pull/4611", "diff_url": "https://github.com/huggingface/datasets/pull/4611.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4611.patch", "merged_at": "2022-07-01T16:48...
4,611
true
codeparrot/github-code failing to load
## Describe the bug codeparrot/github-code fails to load with a `TypeError: get_patterns_in_dataset_repository() missing 1 required positional argument: 'base_path'` ## Steps to reproduce the bug ```python from datasets import load_dataset ``` ## Expected results loaded dataset object ## Actual results `...
https://github.com/huggingface/datasets/issues/4610
[ "I believe the issue is in `codeparrot/github-code`. `base_path` param is missing - https://huggingface.co/datasets/codeparrot/github-code/blob/main/github-code.py#L169\r\n\r\nFunction definition has changed.\r\nhttps://github.com/huggingface/datasets/blob/0e1c629cfb9f9ba124537ba294a0ec451584da5f/src/datasets/data_...
null
4,610
false
librispeech dataset has to download whole subset when specifing the split to use
## Describe the bug librispeech dataset has to download whole subset when specifing the split to use ## Steps to reproduce the bug see below # Sample code to reproduce the bug ``` !pip install datasets from datasets import load_dataset raw_dataset = load_dataset("librispeech_asr", "clean", split="train.100") ...
https://github.com/huggingface/datasets/issues/4609
[ "Hi! You can use streaming to fetch only a subset of the data:\r\n```python\r\nraw_dataset = load_dataset(\"librispeech_asr\", \"clean\", split=\"train.100\", streaming=True)\r\n```\r\nAlso, we plan to make it possible to download a particular split in the non-streaming mode, but this task is not easy due to how ou...
null
4,609
false
Fix xisfile, xgetsize, xisdir, xlistdir in private repo
`xisfile` is working in a private repository when passing a chained URL to a file inside an archive, e.g. `zip://a.txt::https://huggingface/datasets/username/dataset_name/resolve/main/data.zip`. However it's not working when passing a simple file `https://huggingface/datasets/username/dataset_name/resolve/main/data.zip...
https://github.com/huggingface/datasets/pull/4608
[ "_The documentation is not available anymore as the PR was closed or merged._", "Added tests for xisfile, xgetsize, xlistdir and xglob for private repos, and also tests for xwalk that was untested" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4608", "html_url": "https://github.com/huggingface/datasets/pull/4608", "diff_url": "https://github.com/huggingface/datasets/pull/4608.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4608.patch", "merged_at": "2022-07-06T12:34...
4,608
true
Align more metadata with other repo types (models,spaces)
see also associated PR on the `datasets-tagging` Space: https://huggingface.co/spaces/huggingface/datasets-tagging/discussions/2 (to merge after this one is merged)
https://github.com/huggingface/datasets/pull/4607
[ "_The documentation is not available anymore as the PR was closed or merged._", "I just set a default value (None) for the deprecated licenses and languages fields, which should fix most of the CI failures.\r\n\r\nNote that the CI should still be red because you edited many dataset cards and they're still missing...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4607", "html_url": "https://github.com/huggingface/datasets/pull/4607", "diff_url": "https://github.com/huggingface/datasets/pull/4607.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4607.patch", "merged_at": "2022-07-01T11:49...
4,607
true
evaluation result changes after `datasets` version change
## Describe the bug evaluation result changes after `datasets` version change ## Steps to reproduce the bug 1. Train a model on WikiAnn 2. reload the ckpt -> test accuracy becomes same as eval accuracy 3. such behavior is gone after downgrading `datasets` https://colab.research.google.com/drive/1kYz7-aZRGdaya...
https://github.com/huggingface/datasets/issues/4606
[ "Hi! The GH/no-namespace datasets versioning is synced with the version of the `datasets` lib, which means that the `wikiann` script was modified between the two compared versions. In this scenario, you can ensure reproducibility by pinning the script version, which is done by passing `revision=\"x.y.z\"` (e.g. `re...
null
4,606
false
Dataset Viewer issue for boris/gis_filtered
### Link https://huggingface.co/datasets/boris/gis_filtered/viewer/boris--gis_filtered/train ### Description When I try to access this from the website I get this error: Status code: 400 Exception: ClientResponseError Message: 401, message='Unauthorized', url=URL('https://huggingface.co/datase...
https://github.com/huggingface/datasets/issues/4605
[ "Yes, this dataset is \"gated\": you first have to go to https://huggingface.co/datasets/boris/gis_filtered and click \"Access repository\" (if you accept to share your contact information with the repository authors).", "I already did that, it returns error when using streaming", "Oh, sorry, I misread. Looking...
null
4,605
false
Update CI Windows orb
This PR tries to fix recurrent random CI failures on Windows. After 2 runs, it seems to have fixed the issue. Fix #4603.
https://github.com/huggingface/datasets/pull/4604
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4604", "html_url": "https://github.com/huggingface/datasets/pull/4604", "diff_url": "https://github.com/huggingface/datasets/pull/4604.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4604.patch", "merged_at": "2022-06-30T13:22...
4,604
true
CI fails recurrently and randomly on Windows
As reported by @lhoestq, The windows CI is currently flaky: some dependencies like `aiobotocore`, `multiprocess` and `seqeval` sometimes fail to install. In particular it seems that building the wheels fail. Here is an example of logs: ``` Building wheel for seqeval (setup.py): started Running command 'C:\to...
https://github.com/huggingface/datasets/issues/4603
[]
null
4,603
false
Upgrade setuptools in windows CI
The windows CI is currently flaky: some dependencies like aiobotocore, multiprocess and seqeval sometimes fail to install. In particular it seems that building the wheels fail. Here is an example of logs ``` Building wheel for seqeval (setup.py): started Running command 'C:\tools\miniconda3\envs\py37\python.exe...
https://github.com/huggingface/datasets/pull/4602
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4602", "html_url": "https://github.com/huggingface/datasets/pull/4602", "diff_url": "https://github.com/huggingface/datasets/pull/4602.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4602.patch", "merged_at": null }
4,602
true
Upgrade pip in WIN CI
The windows CI is currently flaky: some dependencies like aiobotocore, multiprocess and seqeval sometimes fail to install. In particular it seems that building the wheels fail. Here is an example of logs ``` Building wheel for seqeval (setup.py): started Running command 'C:\tools\miniconda3\envs\py37\python.exe...
https://github.com/huggingface/datasets/pull/4601
[ "_The documentation is not available anymore as the PR was closed or merged._", "It failed terribly" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4601", "html_url": "https://github.com/huggingface/datasets/pull/4601", "diff_url": "https://github.com/huggingface/datasets/pull/4601.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4601.patch", "merged_at": null }
4,601
true
Remove multiple config section
This PR removes docs for a future feature and redirects to #4578 instead. See this [discussion](https://huggingface.slack.com/archives/C034N0A7H09/p1656107063801969) for more details :)
https://github.com/huggingface/datasets/pull/4600
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4600", "html_url": "https://github.com/huggingface/datasets/pull/4600", "diff_url": "https://github.com/huggingface/datasets/pull/4600.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4600.patch", "merged_at": "2022-07-04T17:29...
4,600
true
Smooth-BLEU bug fixed
Hi, the current implementation of smooth-BLEU contains a bug: it smoothes unigrams as well. Consequently, when both the reference and translation consist of totally different tokens, it anyway returns a non-zero value (please see the attached image). This however contradicts the source paper suggesting the smoot...
https://github.com/huggingface/datasets/pull/4599
[ "Thanks @Aktsvigun for your fix.\r\n\r\nHowever, metrics in `datasets` are in deprecation mode:\r\n- #4739\r\n\r\nYou should transfer this PR to the `evaluate` library: https://github.com/huggingface/evaluate\r\n\r\nJust for context, here the link to the PR by @Aktsvigun on tensorflow/nmt:\r\n- https://github.com/t...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4599", "html_url": "https://github.com/huggingface/datasets/pull/4599", "diff_url": "https://github.com/huggingface/datasets/pull/4599.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4599.patch", "merged_at": null }
4,599
true
Host financial_phrasebank data on the Hub
Fix #4597.
https://github.com/huggingface/datasets/pull/4598
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4598", "html_url": "https://github.com/huggingface/datasets/pull/4598", "diff_url": "https://github.com/huggingface/datasets/pull/4598.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4598.patch", "merged_at": "2022-07-01T09:29...
4,598
true
Streaming issue for financial_phrasebank
### Link https://huggingface.co/datasets/financial_phrasebank/viewer/sentences_allagree/train ### Description As reported by a community member using [AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/5#62bc217436d0e5d316a768f0), there seems to be a problem streaming this dat...
https://github.com/huggingface/datasets/issues/4597
[ "cc @huggingface/datasets: it seems like https://www.researchgate.net/ is flaky for datasets hosting (I put the \"hosted-on-google-drive\" tag since it's the same kind of issue I think)", "Let's see if their license allows hosting their data on the Hub.", "License is Creative Commons Attribution-NonCommercial-S...
null
4,597
false
Dataset Viewer issue for universal_dependencies
### Link https://huggingface.co/datasets/universal_dependencies ### Description invalid json response body at https://datasets-server.huggingface.co/splits?dataset=universal_dependencies reason: Unexpected token I in JSON at position 0 ### Owner _No response_
https://github.com/huggingface/datasets/issues/4596
[ "Thanks, looking at it!", "Finally fixed! We updated the dataset viewer and it fixed the issue.\r\n\r\nhttps://huggingface.co/datasets/universal_dependencies/viewer/aqz_tudet/train\r\n\r\n<img width=\"1561\" alt=\"Capture d’écran 2022-09-07 à 13 29 18\" src=\"https://user-images.githubusercontent.com/1676121/18...
null
4,596
false
Dataset Viewer issue with False positive PII redaction
### Link https://huggingface.co/datasets/cakiki/rosetta-code ### Description Hello, I just noticed an entry being redacted that shouldn't have been: `RootMeanSquare@Range[10]` is being displayed as `[email protected][10]` ### Owner _No response_
https://github.com/huggingface/datasets/issues/4595
[ "The value is in the data, it's not an issue with the \"dataset-viewer\".\r\n\r\n<img width=\"1161\" alt=\"Capture d’écran 2022-06-29 à 10 25 51\" src=\"https://user-images.githubusercontent.com/1676121/176389325-4d2a9a7f-1583-45b8-aa7a-960ffaa6a36a.png\">\r\n\r\n Maybe open a PR: https://huggingface.co/datasets/...
null
4,595
false
load_from_disk suggests incorrect fix when used to load DatasetDict
Edit: Please feel free to remove this issue. The problem was not the error message but the fact that the DatasetDict.load_from_disk does not support loading nested splits, i.e. if one of the splits is itself a DatasetDict. If nesting splits is an antipattern, perhaps the load_from_disk function can throw a warning indi...
https://github.com/huggingface/datasets/issues/4594
[]
null
4,594
false
Fix error message when using load_from_disk to load DatasetDict
Issue #4594 Issue: When `datasets.load_from_disk` is wrongly used to load a `DatasetDict`, the error message suggests using `datasets.load_from_disk`, which is the same function that generated the error. Fix: The appropriate function which should be suggested instead is `datasets.dataset_dict.load_from_disk`. Chan...
https://github.com/huggingface/datasets/pull/4593
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4593", "html_url": "https://github.com/huggingface/datasets/pull/4593", "diff_url": "https://github.com/huggingface/datasets/pull/4593.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4593.patch", "merged_at": null }
4,593
true
Issue with jalFaizy/detect_chess_pieces when running datasets-cli test
### Link https://huggingface.co/datasets/jalFaizy/detect_chess_pieces ### Description I am trying to write a appropriate data loader for [a custom dataset](https://huggingface.co/datasets/jalFaizy/detect_chess_pieces) using [this script](https://huggingface.co/datasets/jalFaizy/detect_chess_pieces/blob/main/detect_c...
https://github.com/huggingface/datasets/issues/4592
[ "Hi @faizankshaikh\r\n\r\nPlease note that we have recently launched the Community feature, specifically targeted to create Discussions (about issues/questions/asking-for-help) on each Dataset on the Hub:\r\n- Blog post: https://huggingface.co/blog/community-update\r\n- Docs: https://huggingface.co/docs/hub/reposit...
null
4,592
false
Can't push Images to hub with manual Dataset
## Describe the bug If I create a dataset including an 'Image' feature manually, when pushing to hub decoded images are not pushed, instead it looks for image where image local path is/used to be. This doesn't (at least didn't used to) happen with imagefolder. I want to build dataset manually because it is compli...
https://github.com/huggingface/datasets/issues/4591
[ "Hi, thanks for reporting! This issue stems from the changes introduced in https://github.com/huggingface/datasets/pull/4282 (cc @lhoestq), in which list casts are ignored if they don't change the list type (required to preserve `null` values). And `push_to_hub` does a special cast to embed external image files but...
null
4,591
false
Generalize meta_path json file creation in load.py [#4540]
# What does this PR do? ## Summary *In function `_copy_script_and_other_resources_in_importable_dir`, using string split when generating `meta_path` throws error in the edge case raised in #4540.* ## Additions - ## Changes - Changed meta_path to use `os.path.splitext` instead of using `str.split` to gener...
https://github.com/huggingface/datasets/pull/4590
[ "_The documentation is not available anymore as the PR was closed or merged._", "@albertvillanova, Can you please review this PR for Issue #4540 ", "@lhoestq Thank you for merging the PR . Is there a slack channel for contributing to the datasets library. I would love to work on the library and make meaningfu...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4590", "html_url": "https://github.com/huggingface/datasets/pull/4590", "diff_url": "https://github.com/huggingface/datasets/pull/4590.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4590.patch", "merged_at": "2022-07-07T13:17...
4,590
true
Permission denied: '/home/.cache' when load_dataset with local script
null
https://github.com/huggingface/datasets/issues/4589
[]
null
4,589
false
Host head_qa data on the Hub and fix NonMatchingChecksumError
This PR: - Hosts head_qa data on the Hub instead of Google Drive - Fixes NonMatchingChecksumError Fix https://huggingface.co/datasets/head_qa/discussions/1
https://github.com/huggingface/datasets/pull/4588
[ "_The documentation is not available anymore as the PR was closed or merged._", "Hi @albertvillanova ! Thanks for the fix ;)\r\nCan I safely checkout from this branch to build `datasets` or it is preferable to wait until all CI tests pass?\r\nThanks 🙏 ", "@younesbelkada we have just merged this PR." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4588", "html_url": "https://github.com/huggingface/datasets/pull/4588", "diff_url": "https://github.com/huggingface/datasets/pull/4588.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4588.patch", "merged_at": "2022-07-05T15:49...
4,588
true
Validate new_fingerprint passed by user
Users can pass the dataset fingerprint they want in `map` and other dataset transforms. However the fingerprint is used to name cache files so we need to make sure it doesn't contain bad characters as mentioned in https://github.com/huggingface/datasets/issues/1718, and that it's not too long
https://github.com/huggingface/datasets/pull/4587
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4587", "html_url": "https://github.com/huggingface/datasets/pull/4587", "diff_url": "https://github.com/huggingface/datasets/pull/4587.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4587.patch", "merged_at": "2022-06-28T14:00...
4,587
true
Host pn_summary data on the Hub instead of Google Drive
Fix #4581.
https://github.com/huggingface/datasets/pull/4586
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4586", "html_url": "https://github.com/huggingface/datasets/pull/4586", "diff_url": "https://github.com/huggingface/datasets/pull/4586.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4586.patch", "merged_at": "2022-06-28T14:42...
4,586
true
Host multi_news data on the Hub instead of Google Drive
Host data files of multi_news dataset on the Hub. They were on Google Drive. Fix #4580.
https://github.com/huggingface/datasets/pull/4585
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4585", "html_url": "https://github.com/huggingface/datasets/pull/4585", "diff_url": "https://github.com/huggingface/datasets/pull/4585.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4585.patch", "merged_at": "2022-06-28T14:08...
4,585
true
Add binary classification task IDs
As a precursor to aligning the task IDs in `datasets` and AutoTrain, we need a way to distinguish binary vs multiclass vs multilabel classification. This PR adds binary classification to the task IDs to enable this. Related AutoTrain issue: https://github.com/huggingface/autonlp-backend/issues/597 cc @abhishek...
https://github.com/huggingface/datasets/pull/4584
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4584). All of your documentation changes will be reflected on that endpoint.", "> Awesome thanks ! Can you add it to https://github.com/huggingface/hub-docs/blob/main/js/src/lib/interfaces/Types.ts first please ? This is where ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4584", "html_url": "https://github.com/huggingface/datasets/pull/4584", "diff_url": "https://github.com/huggingface/datasets/pull/4584.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4584.patch", "merged_at": null }
4,584
true
<code> implementation of FLAC support using torchaudio
I had added Audio FLAC support with torchaudio given that Librosa and SoundFile can give problems. Also, FLAC is been used as audio from https://mlcommons.org/en/peoples-speech/
https://github.com/huggingface/datasets/pull/4583
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4583", "html_url": "https://github.com/huggingface/datasets/pull/4583", "diff_url": "https://github.com/huggingface/datasets/pull/4583.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4583.patch", "merged_at": null }
4,583
true
add_column should preserve _indexes
https://github.com/huggingface/datasets/issues/3769#issuecomment-1167146126 doing `.add_column("x",x_data)` also removed any `_indexes` on the dataset, decided this shouldn't be the case. This was because `add_column` was creating a new `Dataset(...)` and wasn't possible to pass indexes on init. with this PR now...
https://github.com/huggingface/datasets/pull/4582
[ "The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4582). All of your documentation changes will be reflected on that endpoint." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4582", "html_url": "https://github.com/huggingface/datasets/pull/4582", "diff_url": "https://github.com/huggingface/datasets/pull/4582.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4582.patch", "merged_at": null }
4,582
true
Dataset Viewer issue for pn_summary
### Link https://huggingface.co/datasets/pn_summary/viewer/1.0.0/validation ### Description Getting an index error on the `validation` and `test` splits: ``` Server error Status code: 400 Exception: IndexError Message: list index out of range ``` ### Owner No
https://github.com/huggingface/datasets/issues/4581
[ "linked to https://github.com/huggingface/datasets/issues/4580#issuecomment-1168373066?", "Note that I refreshed twice this dataset, and I still have (another) error on one of the splits\r\n\r\n```\r\nStatus code: 400\r\nException: ClientResponseError\r\nMessage: 403, message='Forbidden', url=URL('htt...
null
4,581
false
Dataset Viewer issue for multi_news
### Link https://huggingface.co/datasets/multi_news ### Description Not sure what the index error is referring to here: ``` Status code: 400 Exception: IndexError Message: list index out of range ``` ### Owner No
https://github.com/huggingface/datasets/issues/4580
[ "Thanks for reporting, @lewtun.\r\n\r\nI forced the refreshing of the preview and it worked OK for train and validation splits.\r\n\r\nI guess the error has to do with the data files being hosted at Google Drive: this gives errors when requested automatically using scripts.\r\nWe should host them to fix the error. ...
null
4,580
false
Support streaming cfq dataset
Support streaming cfq dataset.
https://github.com/huggingface/datasets/pull/4579
[ "_The documentation is not available anymore as the PR was closed or merged._", "@lhoestq I've been refactoring a little the code:\r\n- Use less RAM by loading only the required samples: only if its index is in the splits file\r\n- Start yielding \"earlier\" in streaming mode: for each `split_idx`:\r\n - either ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4579", "html_url": "https://github.com/huggingface/datasets/pull/4579", "diff_url": "https://github.com/huggingface/datasets/pull/4579.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4579.patch", "merged_at": "2022-07-04T19:23...
4,579
true
[Multi Configs] Use directories to differentiate between subsets/configurations
Currently to define several subsets/configurations of your dataset, you need to use a dataset script. However it would be nice to have a no-code way to to this. For example we could specify different configurations of a dataset (for example, if a dataset contains different languages) with one directory per confi...
https://github.com/huggingface/datasets/issues/4578
[ "I want to be able to create folders in a model.", "How to set new split names, instead of train/test/validation? For example, I have a local dataset, consists of several subsets, named \"A\", \"B\", and \"C\". How can I create a huggingface dataset, with splits A/B/C ?\r\n\r\nThe document in https://huggingface....
null
4,578
false
Add authentication tip to `load_dataset`
Add an authentication tip similar to the one in transformers' `PreTrainedModel.from_pretrained` to `load_dataset`/`load_dataset_builder`.
https://github.com/huggingface/datasets/pull/4577
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4577", "html_url": "https://github.com/huggingface/datasets/pull/4577", "diff_url": "https://github.com/huggingface/datasets/pull/4577.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4577.patch", "merged_at": "2022-07-04T13:01...
4,577
true
Include `metadata.jsonl` in resolved data files
Include `metadata.jsonl` in resolved data files. Fix #4548 @lhoestq ~~https://github.com/huggingface/datasets/commit/d94336d30eef17fc9abc67f67fa1c139661f4e75 adds support for metadata files placed at the root, and https://github.com/huggingface/datasets/commit/4d20618ea7a19bc143ddc5fdff9d79e671fcbb95 accounts fo...
https://github.com/huggingface/datasets/pull/4576
[ "_The documentation is not available anymore as the PR was closed or merged._", "I still don't know if the way we implemented data files resolution could support the metadata.jsonl file without bad side effects for the other packaged builders. In particular here if you have a folder of csv/parquet/whatever files ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4576", "html_url": "https://github.com/huggingface/datasets/pull/4576", "diff_url": "https://github.com/huggingface/datasets/pull/4576.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4576.patch", "merged_at": "2022-06-30T10:15...
4,576
true
Problem about wmt17 zh-en dataset
It seems that in subset casia2015, some samples are like `{'c[hn]':'xxx', 'en': 'aa'}`. So when using `data = load_dataset('wmt17', "zh-en")` to load the wmt17 zh-en dataset, which will raise the exception: ``` Traceback (most recent call last): File "train.py", line 78, in <module> data = load_dataset(args....
https://github.com/huggingface/datasets/issues/4575
[ "Running into the same error with `wmt17/zh-en`, `wmt18/zh-en` and `wmt19/zh-en`.", "@albertvillanova @lhoestq Could you take a look at this issue?", "@winterfell2021 Hi, I wonder where the code you provided should be added. I tried to add them in the `datasets/table.py` in `array_cast` function, however, the '...
null
4,575
false
Support streaming mlsum dataset
Support streaming mlsum dataset. This PR: - pins `fsspec` min version with fixed BlockSizeError: `fsspec[http]>=2021.11.1` - https://github.com/fsspec/filesystem_spec/pull/830 - unpins `s3fs==2021.08.1` to align it with `fsspec` requirement: `s3fs>=2021.11.1` > s3fs 2021.8.1 requires fsspec==2021.08.1 - s...
https://github.com/huggingface/datasets/pull/4574
[ "_The documentation is not available anymore as the PR was closed or merged._", "After unpinning `s3fs` and pinning `fsspec[http]>=2021.11.1`, the CI installs\r\n- `fsspec-2022.1.0`\r\n- `s3fs-0.5.1`\r\n\r\nand raises the following error:\r\n```\r\n ImportError while loading conftest '/home/runner/work/datasets/d...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4574", "html_url": "https://github.com/huggingface/datasets/pull/4574", "diff_url": "https://github.com/huggingface/datasets/pull/4574.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4574.patch", "merged_at": "2022-07-21T12:40...
4,574
true
Fix evaluation metadata for ncbi_disease
This PR fixes the task in the evaluation metadata and removes the metrics info as we've decided this is not a great way to propagate this information downstream.
https://github.com/huggingface/datasets/pull/4573
[ "_The documentation is not available anymore as the PR was closed or merged._", "As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4573", "html_url": "https://github.com/huggingface/datasets/pull/4573", "diff_url": "https://github.com/huggingface/datasets/pull/4573.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4573.patch", "merged_at": null }
4,573
true
Dataset Viewer issue for mlsum
### Link https://huggingface.co/datasets/mlsum/viewer/de/train ### Description There's seems to be a problem with the download / streaming of this dataset: ``` Server error Status code: 400 Exception: BadZipFile Message: File is not a zip file ``` ### Owner No
https://github.com/huggingface/datasets/issues/4572
[ "Thanks for reporting, @lewtun.\r\n\r\nAfter investigation, it seems that the server https://gitlab.lip6.fr does not allow HTTP Range requests.\r\n\r\nWe are trying to find a workaround..." ]
null
4,572
false
Dataset Viewer issue for gsarti/flores_101
### Link https://huggingface.co/datasets/gsarti/flores_101 ### Description It seems like streaming isn't supported for this dataset: ``` Server Error Status code: 400 Exception: NotImplementedError Message: Extraction protocol for TAR archives like 'https://dl.fbaipublicfiles.com/flores101/dataset...
https://github.com/huggingface/datasets/issues/4571
[ "Related to https://github.com/huggingface/datasets/issues/4562#issuecomment-1166911751\r\n\r\nI'll assign @albertvillanova ", "I'm just wondering why we don't have this dataset under:\r\n- the `facebook` namespace\r\n- or the canonical dataset `flores`: why does this only have 2 languages?" ]
null
4,571
false
Dataset sharding non-contiguous?
## Describe the bug I'm not sure if this is a bug; more likely normal behavior but i wanted to double check. Is it normal that `datasets.shard` does not produce chunks that, when concatenated produce the original ordering of the sharded dataset? This might be related to this pull request (https://github.com/huggi...
https://github.com/huggingface/datasets/issues/4570
[ "This was silly; I was sure I'd looked for a `contiguous` argument, and was certain there wasn't one the first time I looked :smile:\r\n\r\nSorry about that.", "Hi! You can pass `contiguous=True` to `.shard()` get contiguous shards. More info on this and the default behavior can be found in the [docs](https://hug...
null
4,570
false
Dataset Viewer issue for sst2
### Link https://huggingface.co/datasets/sst2 ### Description Not sure what is causing this, however it seems that `load_dataset("sst2")` also hangs (even though it downloads the files without problem): ``` Status code: 400 Exception: Exception Message: Give up after 5 attempts with Connectio...
https://github.com/huggingface/datasets/issues/4569
[ "Hi @lewtun, thanks for reporting.\r\n\r\nI have checked locally and refreshed the preview and it seems working smooth now:\r\n```python\r\nIn [8]: ds\r\nOut[8]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['idx', 'sentence', 'label'],\r\n num_rows: 67349\r\n })\r\n validation: Datas...
null
4,569
false
XNLI cache reload is very slow
### Reproduce Using `2.3.3.dev0` `from datasets import load_dataset` `load_dataset("xnli", "en")` Turn off Internet `load_dataset("xnli", "en")` I cancelled the second `load_dataset` eventually cuz it took super long. It would be great to have something to specify e.g. `only_load_from_cache` and avoid the ...
https://github.com/huggingface/datasets/issues/4568
[ "Hi,\r\nCould you tell us how you are running this code?\r\nI tested on my machine (M1 Mac). And it is running fine both on and off internet.\r\n\r\n<img width=\"1033\" alt=\"Screen Shot 2022-07-03 at 1 32 25 AM\" src=\"https://user-images.githubusercontent.com/8711912/177026364-4ad7cedb-e524-4513-97f7-7961bbb34c90...
null
4,568
false
Add evaluation data for amazon_reviews_multi
null
https://github.com/huggingface/datasets/pull/4567
[ "_The documentation is not available anymore as the PR was closed or merged._", "As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4567", "html_url": "https://github.com/huggingface/datasets/pull/4567", "diff_url": "https://github.com/huggingface/datasets/pull/4567.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4567.patch", "merged_at": null }
4,567
true
Document link #load_dataset_enhancing_performance points to nowhere
## Describe the bug A clear and concise description of what the bug is. ![image](https://user-images.githubusercontent.com/11674033/175752806-5b066b92-9d28-4771-9112-5c8606f07741.png) The [load_dataset_enhancing_performance](https://huggingface.co/docs/datasets/v2.3.2/en/package_reference/main_classes#load_dat...
https://github.com/huggingface/datasets/issues/4566
[ "Hi! This is indeed the link the docstring should point to. Are you interested in submitting a PR to fix this?", "https://github.com/huggingface/datasets/blame/master/docs/source/cache.mdx#L93\r\n\r\nThere seems already an anchor here. Somehow it doesn't work. I am not very familiar with how this online documenta...
null
4,566
false
Add UFSC OCPap dataset
## Adding a Dataset - **Name:** UFSC OCPap: Papanicolaou Stained Oral Cytology Dataset (v4) - **Description:** The UFSC OCPap dataset comprises 9,797 labeled images of 1200x1600 pixels acquired from 5 slides of cancer diagnosed and 3 healthy of oral brush samples, from distinct patients. - **Paper:** https://dx.doi....
https://github.com/huggingface/datasets/issues/4565
[ "I will add this directly on the hub (same as #4486)—in https://huggingface.co/lapix" ]
null
4,565
false
Support streaming bookcorpus dataset
Support streaming bookcorpus dataset.
https://github.com/huggingface/datasets/pull/4564
[ "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/4564", "html_url": "https://github.com/huggingface/datasets/pull/4564", "diff_url": "https://github.com/huggingface/datasets/pull/4564.diff", "patch_url": "https://github.com/huggingface/datasets/pull/4564.patch", "merged_at": "2022-07-06T09:23...
4,564
true