title stringlengths 1 290 | body stringlengths 0 228k ⌀ | html_url stringlengths 46 51 | comments list | pull_request dict | number int64 1 5.59k | is_pull_request bool 2
classes |
|---|---|---|---|---|---|---|
Issues with concatenating datasets | ## Describe the bug
It is impossible to concatenate datasets if a feature is sequence of dict in one dataset and a dict of sequence in another. But based on the document, it should be automatically converted.
> A [datasets.Sequence](https://huggingface.co/docs/datasets/v2.3.2/en/package_reference/main_classes#datas... | https://github.com/huggingface/datasets/issues/4666 | [
"Hi! I agree we should improve the features equality checks to account for this particular case. However, your code fails due to `answer_start` having the dtype `int64` instead of `int32` after loading from JSON (it's not possible to embed type precision info into a JSON file; `save_to_disk` does that for arrow fil... | null | 4,666 | false |
Unable to create dataset having Python dataset script only | ## Describe the bug
Hi there,
I'm trying to add the following dataset to Huggingface datasets: https://huggingface.co/datasets/Heriot-WattUniversity/dialog-babi/blob/
I'm trying to do so using the CLI commands but seems that this command generates the wrong `dataset_info.json` file (you can find it in the repo a... | https://github.com/huggingface/datasets/issues/4665 | [
"Hi @aleSuglia, thanks for reporting.\r\n\r\nWe are having a look at it. \r\n\r\nWe transfer this issue to the Community tab of the corresponding Hub dataset: https://huggingface.co/datasets/Heriot-WattUniversity/dialog-babi/discussions"
] | null | 4,665 | false |
Add stanford dog dataset | This PR is for adding dataset, related to issue #4504.
We are adding Stanford dog breed dataset. It is a multi class image classification dataset.
Details can be found here - http://vision.stanford.edu/aditya86/ImageNetDogs/
Tests on dummy data is failing currently, which I am looking into. | https://github.com/huggingface/datasets/pull/4664 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Hi @khushmeeet, thanks for your contribution.\r\n\r\nBut wouldn't it be better to add this dataset to the Hub? \r\n- https://huggingface.co/docs/datasets/share\r\n- https://huggingface.co/docs/datasets/dataset_script",
"Hi @albertv... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4664",
"html_url": "https://github.com/huggingface/datasets/pull/4664",
"diff_url": "https://github.com/huggingface/datasets/pull/4664.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4664.patch",
"merged_at": null
} | 4,664 | true |
Add text decorators | This PR adds some decoration to text about different modalities to make it more obvious separate guides exist for audio, vision, and text. The goal is to make it easier for users to discover these guides!
. ",
"Hi! That's weird. It seems like the error points to the `mkstemp` function, but the official docs state the following:\r\n```\r\nThere are no race condi... | null | 4,661 | false |
Fix _resolve_single_pattern_locally on Windows with multiple drives | Currently, when `_resolve_single_pattern_locally` is called from a different drive than the one in `pattern`, it raises an exception:
```
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
C:\hostedtoolcache\windows\Python\3.6.8\x64\lib\site-packages\datasets\io\parquet.py:35: in __init_... | https://github.com/huggingface/datasets/pull/4660 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Good catch ! Sorry I forgot (again) about windows paths when writing this x)"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4660",
"html_url": "https://github.com/huggingface/datasets/pull/4660",
"diff_url": "https://github.com/huggingface/datasets/pull/4660.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4660.patch",
"merged_at": "2022-07-07T16:52... | 4,660 | true |
Transfer CI to GitHub Actions | This PR transfers CI from CircleCI to GitHub Actions. The implementation in GitHub Actions tries to be as faithful as possible to the implementation in CircleCI and get the same output results (exceptions below).
**IMPORTANT NOTE**: The fast-fail policy (described below) is not finally implemented, so that:
- we c... | https://github.com/huggingface/datasets/pull/4659 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thanks a lot @albertvillanova ! I hope we're finally done with flakiness on windows ^^\r\n\r\nAlso thanks for paying extra attention to billing and avoiding running unnecessary jobs. Though for certain aspects (see my comments), I th... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4659",
"html_url": "https://github.com/huggingface/datasets/pull/4659",
"diff_url": "https://github.com/huggingface/datasets/pull/4659.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4659.patch",
"merged_at": "2022-07-12T11:18... | 4,659 | true |
Transfer CI tests to GitHub Actions | Let's try CI tests using GitHub Actions to see if they are more stable than on CircleCI. | https://github.com/huggingface/datasets/issues/4658 | [] | null | 4,658 | false |
Add SQuAD2.0 Dataset | ## Adding a Dataset
- **Name:** *SQuAD2.0*
- **Description:** *Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading ... | https://github.com/huggingface/datasets/issues/4657 | [
"Hey, It's already present [here](https://huggingface.co/datasets/squad_v2) ",
"Hi! This dataset is indeed already available on the Hub. Closing."
] | null | 4,657 | false |
Add Amazon-QA Dataset | ## Adding a Dataset
- **Name:** *Amazon-QA*
- **Description:** *The dataset is .jsonl format, where each line in the file is a json string that corresponds to a question, existing answers to the question and the extracted review snippets (relevant to the question).*
- **Paper:** *https://github.com/amazonqa/amazonqa... | https://github.com/huggingface/datasets/issues/4656 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/Amazon-QA)."
] | null | 4,656 | false |
Simple Wikipedia | ## Adding a Dataset
- **Name:** *Simple Wikipedia*
- **Description:** *Two different versions of the data set now exist. Both were generated by aligning Simple English Wikipedia and English Wikipedia. A complete description of the extraction process can be found in "Simple English Wikipedia: A New Simplification Task... | https://github.com/huggingface/datasets/issues/4655 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/simple-wiki)."
] | null | 4,655 | false |
Add Quora Question Triplets Dataset | ## Adding a Dataset
- **Name:** *Quora Question Triplets*
- **Description:** *This dataset consists of over 400,000 lines of potential question duplicate pairs. Each line contains IDs for each question in the pair, the full text for each question, and a binary value that indicates whether the line truly contains a du... | https://github.com/huggingface/datasets/issues/4654 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/QQP_triplets)."
] | null | 4,654 | false |
Add Altlex dataset | ## Adding a Dataset
- **Name:** *Altlex*
- **Description:** *Git repository for software associated with the 2016 ACL paper "Identifying Causal Relations Using Parallel Wikipedia Articles.”*
- **Paper:** *https://aclanthology.org/P16-1135.pdf*
- **Data:** *https://huggingface.co/datasets/sentence-transformers/embed... | https://github.com/huggingface/datasets/issues/4653 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/altlex)."
] | null | 4,653 | false |
Add Sentence Compression Dataset | ## Adding a Dataset
- **Name:** *Sentence Compression*
- **Description:** *Large corpus of uncompressed and compressed sentences from news articles.*
- **Paper:** *https://www.aclweb.org/anthology/D13-1155/*
- **Data:** *https://github.com/google-research-datasets/sentence-compression/tree/master/data*
- **Motivat... | https://github.com/huggingface/datasets/issues/4652 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/sentence-compression)."
] | null | 4,652 | false |
Add Flickr 30k Dataset | ## Adding a Dataset
- **Name:** *Flickr 30k*
- **Description:** *To produce the denotation graph, we have created an image caption corpus consisting of 158,915 crowd-sourced captions describing 31,783 images. This is an extension of our previous Flickr 8k Dataset. The new images and captions focus on people involved ... | https://github.com/huggingface/datasets/issues/4651 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/flickr30k-captions)."
] | null | 4,651 | false |
Add SPECTER dataset | ## Adding a Dataset
- **Name:** *SPECTER*
- **Description:** *SPECTER: Document-level Representation Learning using Citation-informed Transformers*
- **Paper:** *https://doi.org/10.18653/v1/2020.acl-main.207*
- **Data:** *https://huggingface.co/datasets/sentence-transformers/embedding-training-data/resolve/main/spe... | https://github.com/huggingface/datasets/issues/4650 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/SPECTER)"
] | null | 4,650 | false |
Add PAQ dataset | ## Adding a Dataset
- **Name:** *PAQ*
- **Description:** *This repository contains code and models to support the research paper PAQ: 65 Million Probably-Asked Questions and What You Can Do With Them*
- **Paper:** *https://arxiv.org/abs/2102.07033*
- **Data:** *https://huggingface.co/datasets/sentence-transformers/... | https://github.com/huggingface/datasets/issues/4649 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/PAQ_pairs)"
] | null | 4,649 | false |
Add WikiAnswers dataset | ## Adding a Dataset
- **Name:** *WikiAnswers*
- **Description:** *The WikiAnswers corpus contains clusters of questions tagged by WikiAnswers users as paraphrases. Each cluster optionally contains an answer provided by WikiAnswers users.*
- **Paper:** *https://dl.acm.org/doi/10.1145/2623330.2623677*
- **Data:** *ht... | https://github.com/huggingface/datasets/issues/4648 | [
"uploaded dataset [here](https://huggingface.co/datasets/embedding-data/WikiAnswers)"
] | null | 4,648 | false |
Add Reddit dataset | ## Adding a Dataset
- **Name:** *Reddit comments (2015-2018)*
- **Description:** *Reddit is an American social news aggregation website, where users can post links, and take part in discussions on these posts. These threaded discussions provide a large corpus, which is converted into a conversational dataset using th... | https://github.com/huggingface/datasets/issues/4647 | [] | null | 4,647 | false |
Set HF_SCRIPTS_VERSION to main | After renaming "master" to "main", the CI fails with
```
AssertionError: 'https://raw.githubusercontent.com/huggingface/datasets/main/datasets/_dummy/_dummy.py' not found in "Couldn't find a dataset script at /home/circleci/datasets/_dummy/_dummy.py or any data file in the same directory. Couldn't find '_dummy' on th... | https://github.com/huggingface/datasets/pull/4645 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4645",
"html_url": "https://github.com/huggingface/datasets/pull/4645",
"diff_url": "https://github.com/huggingface/datasets/pull/4645.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4645.patch",
"merged_at": "2022-07-06T15:45... | 4,645 | true |
[Minor fix] Typo correction | recieve -> receive | https://github.com/huggingface/datasets/pull/4644 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4644",
"html_url": "https://github.com/huggingface/datasets/pull/4644",
"diff_url": "https://github.com/huggingface/datasets/pull/4644.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4644.patch",
"merged_at": "2022-07-06T15:45... | 4,644 | true |
Rename master to main | This PR renames mentions of "master" by "main" in the code base for several cases:
- set the default dataset script version to "main" if the local installation of `datasets` is a dev installation
- update URLs to this github repository to use "main"
- update the DVC benchmark
- update the github workflows
- update... | https://github.com/huggingface/datasets/pull/4643 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"All the mentions I found on google were simple URLs that will be redirected, so it's fine. I also checked the spaces and we should be good:\r\n- dalle-mini used to install the master branch but [it's no longer the case](https://huggi... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4643",
"html_url": "https://github.com/huggingface/datasets/pull/4643",
"diff_url": "https://github.com/huggingface/datasets/pull/4643.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4643.patch",
"merged_at": "2022-07-06T15:25... | 4,643 | true |
Streaming issue for ccdv/pubmed-summarization | ### Link
https://huggingface.co/datasets/ccdv/pubmed-summarization
### Description
This was reported by a [user of AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/7). It seems like streaming doesn't work due to the way the dataset loading script is defined?
```
Status c... | https://github.com/huggingface/datasets/issues/4642 | [
"Thanks for reporting @lewtun.\r\n\r\nI confirm there is an issue with streaming: it does not stream locally. ",
"Oh, after investigation, the source of the issue is in the Hub dataset loading script.\r\n\r\nI'm opening a PR on the Hub dataset.",
"I've opened a PR on their Hub dataset to support streaming: http... | null | 4,642 | false |
Dataset Viewer issue for kmfoda/booksum | ### Link
https://huggingface.co/datasets/kmfoda/booksum
### Description
A [user of AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/9) discovered this dataset cannot be streamed due to:
```
Status code: 400
Exception: ClientResponseError
Message: 401, messa... | https://github.com/huggingface/datasets/issues/4641 | [
"Thanks for reporting, @lewtun.\r\n\r\nIt works locally in streaming mode:\r\n```\r\n{'bid': 27681,\r\n 'is_aggregate': True,\r\n 'source': 'cliffnotes',\r\n 'chapter_path': 'all_chapterized_books/27681-chapters/chapters_1_to_2.txt',\r\n 'summary_path': 'finished_summaries/cliffnotes/The Last of the Mohicans/sectio... | null | 4,641 | false |
Support all split in streaming mode | Fix #4637. | https://github.com/huggingface/datasets/pull/4640 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4640). All of your documentation changes will be reflected on that endpoint."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4640",
"html_url": "https://github.com/huggingface/datasets/pull/4640",
"diff_url": "https://github.com/huggingface/datasets/pull/4640.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4640.patch",
"merged_at": null
} | 4,640 | true |
Add HaGRID -- HAnd Gesture Recognition Image Dataset | ## Adding a Dataset
- **Name:** HaGRID -- HAnd Gesture Recognition Image Dataset
- **Description:** We introduce a large image dataset HaGRID (HAnd Gesture Recognition Image Dataset) for hand gesture recognition (HGR) systems. You can use it for image classification or image detection tasks. Proposed dataset allows t... | https://github.com/huggingface/datasets/issues/4639 | [] | null | 4,639 | false |
The speechocean762 dataset | [speechocean762](https://www.openslr.org/101/) is a non-native English corpus for pronunciation scoring tasks. It is free for both commercial and non-commercial use.
I believe it will be easier to use if it could be available on Hugging Face. | https://github.com/huggingface/datasets/pull/4638 | [
"CircleCL reported two errors, but I didn't find the reason. The error message:\r\n```\r\n_________________ ERROR collecting tests/test_dataset_cards.py _________________\r\ntests/test_dataset_cards.py:53: in <module>\r\n @pytest.mark.parametrize(\"dataset_name\", get_changed_datasets(repo_path))\r\ntests/test_d... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4638",
"html_url": "https://github.com/huggingface/datasets/pull/4638",
"diff_url": "https://github.com/huggingface/datasets/pull/4638.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4638.patch",
"merged_at": null
} | 4,638 | true |
The "all" split breaks streaming | ## Describe the bug
Not sure if this is a bug or just the way streaming works, but setting `streaming=True` did not work when setting `split="all"`
## Steps to reproduce the bug
The following works:
```python
ds = load_dataset('super_glue', 'wsc.fixed', split='all')
```
The following throws `ValueError: Bad ... | https://github.com/huggingface/datasets/issues/4637 | [
"Thanks for reporting @cakiki.\r\n\r\nYes, this is a bug. We are investigating it.",
"@albertvillanova Nice! Let me know if it's something I can fix my self; would love to contribtue!",
"@cakiki I was working on this but if you would like to contribute, go ahead. I will close my PR. ;)\r\n\r\nFor the moment I j... | null | 4,637 | false |
Add info in docs about behavior of download_config.num_proc | **Is your feature request related to a problem? Please describe.**
I went to override `download_config.num_proc` and was confused about what was happening under the hood. It would be nice to have the behavior documented a bit better so folks know what's happening when they use it.
**Describe the solution you'd li... | https://github.com/huggingface/datasets/issues/4636 | [] | null | 4,636 | false |
Dataset Viewer issue for vadis/sv-ident | ### Link
https://huggingface.co/datasets/vadis/sv-ident/viewer/default/validation
### Description
Error message when loading validation split in the viewer:
```
Status code: 400
Exception: Status400Error
Message: The split cache is empty.
```
### Owner
_No response_ | https://github.com/huggingface/datasets/issues/4635 | [
"Thanks for reporting, @e-tornike \r\n\r\nSome context:\r\n- #4527 \r\n\r\nThe dataset loads locally in streaming mode:\r\n```python\r\nIn [2]: from datasets import load_dataset; ds = load_dataset(\"vadis/sv-ident\", split=\"validation\", streaming=True); item = next(iter(ds)); item\r\nUsing custom data configurati... | null | 4,635 | false |
Can't load the Hausa audio dataset | common_voice_train = load_dataset("common_voice", "ha", split="train+validation") | https://github.com/huggingface/datasets/issues/4634 | [
"Could you provide the error details. It is difficult to debug otherwise. Also try other config. `ha` is not a valid."
] | null | 4,634 | false |
[data_files] Only match separated split names | As reported in https://github.com/huggingface/datasets/issues/4477, the current pattern matching to infer which file goes into which split is too permissive. For example a file "contest.py" would be considered part of a test split (it contains "test") and "seqeval.py" as well (it contains "eval").
In this PR I made ... | https://github.com/huggingface/datasets/pull/4633 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I ran a script to find affected datasets (just did it on non-private non-gated). Adding \"testing\" and \"evaluation\" fixes all of of them except one:\r\n- projecte-aina/cat_manynames:\thuman_annotated_testset.tsv\r\n\r\nLet me open... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4633",
"html_url": "https://github.com/huggingface/datasets/pull/4633",
"diff_url": "https://github.com/huggingface/datasets/pull/4633.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4633.patch",
"merged_at": "2022-07-18T13:07... | 4,633 | true |
'sort' method sorts one column only | The 'sort' method changes the order of one column only (the one defined by the argument 'column'), thus creating a mismatch between a sample fields. I would expect it to change the order of the samples as a whole, based on the 'column' order. | https://github.com/huggingface/datasets/issues/4632 | [
"Hi ! `ds.sort()` does sort the full dataset, not just one column:\r\n```python\r\nfrom datasets import *\r\n\r\nds = Dataset.from_dict({\"foo\": [3, 2, 1], \"bar\": [\"c\", \"b\", \"a\"]})\r\nprint(d.sort(\"foo\").to_pandas()\r\n# foo bar\r\n# 0 1 a\r\n# 1 2 b\r\n# 2 3 c\r\n```\r\n\r\nWhat made y... | null | 4,632 | false |
Update WinoBias README | I'm adding some information about Winobias that I got from the paper :smile:
I think this makes it a bit clearer! | https://github.com/huggingface/datasets/pull/4631 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4631",
"html_url": "https://github.com/huggingface/datasets/pull/4631",
"diff_url": "https://github.com/huggingface/datasets/pull/4631.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4631.patch",
"merged_at": "2022-07-07T13:11... | 4,631 | true |
fix(dataset_wrappers): Fixes access to fsspec.asyn in torch_iterable_dataset.py. | Fix #4612.
Apparently, newest `fsspec` versions do not allow access to attribute-based modules if they are not imported, such as `fsspec.async`.
Thus, @mariosasko suggested to add the missing part to the module import to allow for its access. | https://github.com/huggingface/datasets/pull/4630 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4630",
"html_url": "https://github.com/huggingface/datasets/pull/4630",
"diff_url": "https://github.com/huggingface/datasets/pull/4630.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4630.patch",
"merged_at": "2022-07-05T15:08... | 4,630 | true |
Rename repo default branch to main | Rename repository default branch to `main` (instead of current `master`).
Once renamed, users will have to manually update their local repos:
- [ ] Upstream:
```
git branch -m master main
git fetch upstream main
git branch -u upstream/main main
git remote set-head upstream -a
```
- [ ] Origin... | https://github.com/huggingface/datasets/issues/4629 | [] | null | 4,629 | false |
Fix time type `_arrow_to_datasets_dtype` conversion | Fix #4620
The issue stems from the fact that `pa.array([time_data]).type` returns `DataType(time64[unit])`, which doesn't expose the `unit` attribute, instead of `Time64Type(time64[unit])`. I believe this is a bug in PyArrow. Luckily, the both types have the same `str()`, so in this PR I call `pa.type_for_alias(str(... | https://github.com/huggingface/datasets/pull/4628 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4628",
"html_url": "https://github.com/huggingface/datasets/pull/4628",
"diff_url": "https://github.com/huggingface/datasets/pull/4628.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4628.patch",
"merged_at": "2022-07-07T13:57... | 4,628 | true |
fixed duplicate calculation of spearmanr function in metrics wrapper. | During _compute, the scipy.stats spearmanr function was called twice, redundantly, once for calculating the score and once for calculating the p-value, under the conditional branch where return_pvalue=True. I adjusted the _compute function to execute the spearmanr function once, store the results tuple in a temporary v... | https://github.com/huggingface/datasets/pull/4627 | [
"Great, can open a PR in `evaluate` as well to optimize this.\r\n\r\nRelatedly, I wanted to add a new metric, Kendall Tau (https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.kendalltau.html). If I were to open a PR with the wrapper, description, citation, docstrings, readme, etc. would it make more se... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4627",
"html_url": "https://github.com/huggingface/datasets/pull/4627",
"diff_url": "https://github.com/huggingface/datasets/pull/4627.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4627.patch",
"merged_at": "2022-07-07T12:41... | 4,627 | true |
Add non-commercial licensing info for datasets for which we removed tags | We removed several YAML tags saying that certain datasets can't be used for commercial purposes: https://github.com/huggingface/datasets/pull/4613#discussion_r911919753
Reason for this is that we only allow tags that are part of our [supported list of licenses](https://github.com/huggingface/datasets/blob/84fc3ad73c... | https://github.com/huggingface/datasets/issues/4626 | [
"yep plus `license_details` also makes sense for this IMO"
] | null | 4,626 | false |
Unpack `dl_manager.iter_files` to allow parallization | Iterate over data files outside `dl_manager.iter_files` to allow parallelization in streaming mode.
(The issue reported [here](https://discuss.huggingface.co/t/dataset-only-have-n-shard-1-when-has-multiple-shards-in-repo/19887))
PS: Another option would be to override `FilesIterable.__getitem__` to make it indexa... | https://github.com/huggingface/datasets/pull/4625 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Cool thanks ! Yup it sounds like the right solution.\r\n\r\nIt looks like `_generate_tables` needs to be updated as well to fix the CI"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4625",
"html_url": "https://github.com/huggingface/datasets/pull/4625",
"diff_url": "https://github.com/huggingface/datasets/pull/4625.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4625.patch",
"merged_at": "2022-07-05T11:00... | 4,625 | true |
Remove all paperswithcode_id: null | On the Hub there is a validation error on the `paperswithcode_id` tag when the value is `null`:
<img width="686" alt="image" src="https://user-images.githubusercontent.com/42851186/177151825-93d341c5-25bd-41ab-96c2-c0b516d51c68.png">
We've been using `null` to specify that we checked on pwc but the dataset doesn'... | https://github.com/huggingface/datasets/pull/4624 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"> We've been using `null` to specify that we checked on pwc but the dataset doesn't exist there.\r\n\r\n@lhoestq maybe it's better to accept it on the Hub side then? Let me know if you want us to do it Hub-side",
"Yup it's maybe be... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4624",
"html_url": "https://github.com/huggingface/datasets/pull/4624",
"diff_url": "https://github.com/huggingface/datasets/pull/4624.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4624.patch",
"merged_at": null
} | 4,624 | true |
Loading MNIST as Pytorch Dataset | ## Describe the bug
Conversion of MNIST dataset to pytorch fails with bug
## Steps to reproduce the bug
```python
from datasets import load_dataset
dataset = load_dataset("mnist", split="train")
dataset.set_format('torch')
dataset[0]
print()
```
## Expected results
Expect to see torch tensors image and l... | https://github.com/huggingface/datasets/issues/4623 | [
"Hi ! We haven't implemented the conversion from images data to PyTorch tensors yet I think\r\n\r\ncc @mariosasko ",
"So I understand:\r\n\r\nset_format() does not properly do the conversion to pytorch tensors from PIL images.\r\n\r\nSo that someone who stumbles on this can use the package:\r\n\r\n```python\r\nda... | null | 4,623 | false |
Fix ImageFolder with parameters drop_metadata=True and drop_labels=False (when metadata.jsonl is present) | Will fix #4621
ImageFolder raises `KeyError: 'label'` with params `drop_metadata=True` and `drop_labels=False` (if there is at least one metadata.jsonl file a data directory). This happens because metadata files are collected inside `analyze()` function regardless of `drop_metadata` value. And then the following co... | https://github.com/huggingface/datasets/pull/4622 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@lhoestq @mariosasko pls take a look at https://github.com/huggingface/datasets/pull/4622/commits/769e4c046a5bd5e3a4dbd09cfad1f4cf60677869. I modified `_generate_examples()` according to the same logic too: removed checking if `metad... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4622",
"html_url": "https://github.com/huggingface/datasets/pull/4622",
"diff_url": "https://github.com/huggingface/datasets/pull/4622.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4622.patch",
"merged_at": "2022-07-15T14:24... | 4,622 | true |
ImageFolder raises an error with parameters drop_metadata=True and drop_labels=False when metadata.jsonl is present | ## Describe the bug
If you pass `drop_metadata=True` and `drop_labels=False` when a `data_dir` contains at least one `matadata.jsonl` file, you will get a KeyError. This is probably not a very useful case but we shouldn't get an error anyway. Asking users to move metadata files manually outside `data_dir` or pass fe... | https://github.com/huggingface/datasets/issues/4621 | [] | null | 4,621 | false |
Data type is not recognized when using datetime.time | ## Describe the bug
Creating a dataset from a pandas dataframe with `datetime.time` format generates an error.
## Steps to reproduce the bug
```python
import pandas as pd
from datetime import time
from datasets import Dataset
df = pd.DataFrame({"feature_name": [time(1, 1, 1)]})
dataset = Dataset.from_pandas... | https://github.com/huggingface/datasets/issues/4620 | [
"cc @mariosasko ",
"Hi, thanks for reporting! I'm investigating the issue."
] | null | 4,620 | false |
np arrays get turned into native lists | ## Describe the bug
When attaching an `np.array` field, it seems that it automatically gets turned into a list (see below). Why is this happening? Could it lose precision? Is there a way to make sure this doesn't happen?
## Steps to reproduce the bug
```python
>>> import datasets, numpy as np
>>> dataset = datas... | https://github.com/huggingface/datasets/issues/4619 | [
"If you add the line `dataset2.set_format('np')` before calling `dataset2[0]['tmp']` it should return `np.ndarray`.\r\nI believe internally it will not store it as a list, it is only returning a list when you index it.\r\n\r\n```\r\nIn [1]: import datasets, numpy as np\r\nIn [2]: dataset = datasets.load_dataset(\"g... | null | 4,619 | false |
contribute data loading for object detection datasets with yolo data format | **Is your feature request related to a problem? Please describe.**
At the moment, HF datasets loads [image classification datasets](https://huggingface.co/docs/datasets/image_process) out-of-the-box. There could be a data loader for loading standard object detection datasets ([original discussion here](https://hugging... | https://github.com/huggingface/datasets/issues/4618 | [
"Hi! The `imagefolder` script is already quite complex, so a standalone script sounds better. Also, I suggest we create an org on the Hub (e.g. `hf-loaders`) and store such scripts there for easier maintenance rather than having them as packaged modules (IMO only very generic loaders should be packaged). WDYT @lhoe... | null | 4,618 | false |
Fix `embed_storage` on features inside lists/sequences | Add a dedicated function for embed_storage to always preserve the embedded/casted arrays (and to have more control over `embed_storage` in general).
Fix #4591
~~(Waiting for #4608 to be merged to mark this PR as ready for review - required for fixing `xgetsize` in private repos)~~ Done! | https://github.com/huggingface/datasets/pull/4615 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4615",
"html_url": "https://github.com/huggingface/datasets/pull/4615",
"diff_url": "https://github.com/huggingface/datasets/pull/4615.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4615.patch",
"merged_at": "2022-07-08T12:01... | 4,615 | true |
Ensure ConcatenationTable.cast uses target_schema metadata | Currently, `ConcatenationTable.cast` does not use target_schema metadata when casting subtables. This causes an issue when using cast_column and the underlying table is a ConcatenationTable.
Code example of where issue arrises:
```
from datasets import Dataset, Image
column1 = [0, 1]
image_paths = ['/images/im... | https://github.com/huggingface/datasets/pull/4614 | [
"Hi @lhoestq, Thanks for the detailed comment. I've tested the suggested approach and can confirm it works for the testcase outlined above! The PR is updated with the changes.",
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4614",
"html_url": "https://github.com/huggingface/datasets/pull/4614",
"diff_url": "https://github.com/huggingface/datasets/pull/4614.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4614.patch",
"merged_at": "2022-07-19T13:36... | 4,614 | true |
Align/fix license metadata info | fix bad "other-*" licenses and add the corresponding "license_details" when relevant | https://github.com/huggingface/datasets/pull/4613 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Thank you thank you! Let's merge and pray? 😱 ",
"I just need to add `license_details` to the validator and yup we can merge"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4613",
"html_url": "https://github.com/huggingface/datasets/pull/4613",
"diff_url": "https://github.com/huggingface/datasets/pull/4613.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4613.patch",
"merged_at": "2022-07-01T12:42... | 4,613 | true |
Release 2.3.0 broke custom iterable datasets | ## Describe the bug
Trying to iterate examples from custom iterable dataset fails to bug introduced in `torch_iterable_dataset.py` since the release of 2.3.0.
## Steps to reproduce the bug
```python
next(iter(custom_iterable_dataset))
```
## Expected results
`next(iter(custom_iterable_dataset))` should retu... | https://github.com/huggingface/datasets/issues/4612 | [
"Apparently, `fsspec` does not allow access to attribute-based modules anymore, such as `fsspec.async`.\r\n\r\nHowever, this is a fairly simple fix:\r\n- Change the import to: `from fsspec import asyn`;\r\n- Change line 18 to: `asyn.iothread[0] = None`;\r\n- Change line 19 to `asyn.loop[0] = None`.",
"Hi! I think... | null | 4,612 | false |
Preserve member order by MockDownloadManager.iter_archive | Currently, `MockDownloadManager.iter_archive` yields paths to archive members in an order given by `path.rglob("*")`, which migh not be the same order as in the original archive.
See issue in:
- https://github.com/huggingface/datasets/pull/4579#issuecomment-1172135027
This PR fixes the order of the members yield... | https://github.com/huggingface/datasets/pull/4611 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4611",
"html_url": "https://github.com/huggingface/datasets/pull/4611",
"diff_url": "https://github.com/huggingface/datasets/pull/4611.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4611.patch",
"merged_at": "2022-07-01T16:48... | 4,611 | true |
codeparrot/github-code failing to load | ## Describe the bug
codeparrot/github-code fails to load with a `TypeError: get_patterns_in_dataset_repository() missing 1 required positional argument: 'base_path'`
## Steps to reproduce the bug
```python
from datasets import load_dataset
```
## Expected results
loaded dataset object
## Actual results
`... | https://github.com/huggingface/datasets/issues/4610 | [
"I believe the issue is in `codeparrot/github-code`. `base_path` param is missing - https://huggingface.co/datasets/codeparrot/github-code/blob/main/github-code.py#L169\r\n\r\nFunction definition has changed.\r\nhttps://github.com/huggingface/datasets/blob/0e1c629cfb9f9ba124537ba294a0ec451584da5f/src/datasets/data_... | null | 4,610 | false |
librispeech dataset has to download whole subset when specifing the split to use | ## Describe the bug
librispeech dataset has to download whole subset when specifing the split to use
## Steps to reproduce the bug
see below
# Sample code to reproduce the bug
```
!pip install datasets
from datasets import load_dataset
raw_dataset = load_dataset("librispeech_asr", "clean", split="train.100")
... | https://github.com/huggingface/datasets/issues/4609 | [
"Hi! You can use streaming to fetch only a subset of the data:\r\n```python\r\nraw_dataset = load_dataset(\"librispeech_asr\", \"clean\", split=\"train.100\", streaming=True)\r\n```\r\nAlso, we plan to make it possible to download a particular split in the non-streaming mode, but this task is not easy due to how ou... | null | 4,609 | false |
Fix xisfile, xgetsize, xisdir, xlistdir in private repo | `xisfile` is working in a private repository when passing a chained URL to a file inside an archive, e.g. `zip://a.txt::https://huggingface/datasets/username/dataset_name/resolve/main/data.zip`. However it's not working when passing a simple file `https://huggingface/datasets/username/dataset_name/resolve/main/data.zip... | https://github.com/huggingface/datasets/pull/4608 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Added tests for xisfile, xgetsize, xlistdir and xglob for private repos, and also tests for xwalk that was untested"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4608",
"html_url": "https://github.com/huggingface/datasets/pull/4608",
"diff_url": "https://github.com/huggingface/datasets/pull/4608.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4608.patch",
"merged_at": "2022-07-06T12:34... | 4,608 | true |
Align more metadata with other repo types (models,spaces) | see also associated PR on the `datasets-tagging` Space: https://huggingface.co/spaces/huggingface/datasets-tagging/discussions/2 (to merge after this one is merged) | https://github.com/huggingface/datasets/pull/4607 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I just set a default value (None) for the deprecated licenses and languages fields, which should fix most of the CI failures.\r\n\r\nNote that the CI should still be red because you edited many dataset cards and they're still missing... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4607",
"html_url": "https://github.com/huggingface/datasets/pull/4607",
"diff_url": "https://github.com/huggingface/datasets/pull/4607.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4607.patch",
"merged_at": "2022-07-01T11:49... | 4,607 | true |
evaluation result changes after `datasets` version change | ## Describe the bug
evaluation result changes after `datasets` version change
## Steps to reproduce the bug
1. Train a model on WikiAnn
2. reload the ckpt -> test accuracy becomes same as eval accuracy
3. such behavior is gone after downgrading `datasets`
https://colab.research.google.com/drive/1kYz7-aZRGdaya... | https://github.com/huggingface/datasets/issues/4606 | [
"Hi! The GH/no-namespace datasets versioning is synced with the version of the `datasets` lib, which means that the `wikiann` script was modified between the two compared versions. In this scenario, you can ensure reproducibility by pinning the script version, which is done by passing `revision=\"x.y.z\"` (e.g. `re... | null | 4,606 | false |
Dataset Viewer issue for boris/gis_filtered | ### Link
https://huggingface.co/datasets/boris/gis_filtered/viewer/boris--gis_filtered/train
### Description
When I try to access this from the website I get this error:
Status code: 400
Exception: ClientResponseError
Message: 401, message='Unauthorized', url=URL('https://huggingface.co/datase... | https://github.com/huggingface/datasets/issues/4605 | [
"Yes, this dataset is \"gated\": you first have to go to https://huggingface.co/datasets/boris/gis_filtered and click \"Access repository\" (if you accept to share your contact information with the repository authors).",
"I already did that, it returns error when using streaming",
"Oh, sorry, I misread. Looking... | null | 4,605 | false |
Update CI Windows orb | This PR tries to fix recurrent random CI failures on Windows.
After 2 runs, it seems to have fixed the issue.
Fix #4603. | https://github.com/huggingface/datasets/pull/4604 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4604",
"html_url": "https://github.com/huggingface/datasets/pull/4604",
"diff_url": "https://github.com/huggingface/datasets/pull/4604.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4604.patch",
"merged_at": "2022-06-30T13:22... | 4,604 | true |
CI fails recurrently and randomly on Windows | As reported by @lhoestq,
The windows CI is currently flaky: some dependencies like `aiobotocore`, `multiprocess` and `seqeval` sometimes fail to install.
In particular it seems that building the wheels fail. Here is an example of logs:
```
Building wheel for seqeval (setup.py): started
Running command 'C:\to... | https://github.com/huggingface/datasets/issues/4603 | [] | null | 4,603 | false |
Upgrade setuptools in windows CI | The windows CI is currently flaky: some dependencies like aiobotocore, multiprocess and seqeval sometimes fail to install.
In particular it seems that building the wheels fail. Here is an example of logs
```
Building wheel for seqeval (setup.py): started
Running command 'C:\tools\miniconda3\envs\py37\python.exe... | https://github.com/huggingface/datasets/pull/4602 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4602",
"html_url": "https://github.com/huggingface/datasets/pull/4602",
"diff_url": "https://github.com/huggingface/datasets/pull/4602.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4602.patch",
"merged_at": null
} | 4,602 | true |
Upgrade pip in WIN CI | The windows CI is currently flaky: some dependencies like aiobotocore, multiprocess and seqeval sometimes fail to install.
In particular it seems that building the wheels fail. Here is an example of logs
```
Building wheel for seqeval (setup.py): started
Running command 'C:\tools\miniconda3\envs\py37\python.exe... | https://github.com/huggingface/datasets/pull/4601 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"It failed terribly"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4601",
"html_url": "https://github.com/huggingface/datasets/pull/4601",
"diff_url": "https://github.com/huggingface/datasets/pull/4601.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4601.patch",
"merged_at": null
} | 4,601 | true |
Remove multiple config section | This PR removes docs for a future feature and redirects to #4578 instead. See this [discussion](https://huggingface.slack.com/archives/C034N0A7H09/p1656107063801969) for more details :) | https://github.com/huggingface/datasets/pull/4600 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4600",
"html_url": "https://github.com/huggingface/datasets/pull/4600",
"diff_url": "https://github.com/huggingface/datasets/pull/4600.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4600.patch",
"merged_at": "2022-07-04T17:29... | 4,600 | true |
Smooth-BLEU bug fixed | Hi,
the current implementation of smooth-BLEU contains a bug: it smoothes unigrams as well. Consequently, when both the reference and translation consist of totally different tokens, it anyway returns a non-zero value (please see the attached image).
This however contradicts the source paper suggesting the smoot... | https://github.com/huggingface/datasets/pull/4599 | [
"Thanks @Aktsvigun for your fix.\r\n\r\nHowever, metrics in `datasets` are in deprecation mode:\r\n- #4739\r\n\r\nYou should transfer this PR to the `evaluate` library: https://github.com/huggingface/evaluate\r\n\r\nJust for context, here the link to the PR by @Aktsvigun on tensorflow/nmt:\r\n- https://github.com/t... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4599",
"html_url": "https://github.com/huggingface/datasets/pull/4599",
"diff_url": "https://github.com/huggingface/datasets/pull/4599.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4599.patch",
"merged_at": null
} | 4,599 | true |
Host financial_phrasebank data on the Hub |
Fix #4597. | https://github.com/huggingface/datasets/pull/4598 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4598",
"html_url": "https://github.com/huggingface/datasets/pull/4598",
"diff_url": "https://github.com/huggingface/datasets/pull/4598.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4598.patch",
"merged_at": "2022-07-01T09:29... | 4,598 | true |
Streaming issue for financial_phrasebank | ### Link
https://huggingface.co/datasets/financial_phrasebank/viewer/sentences_allagree/train
### Description
As reported by a community member using [AutoTrain Evaluate](https://huggingface.co/spaces/autoevaluate/model-evaluator/discussions/5#62bc217436d0e5d316a768f0), there seems to be a problem streaming this dat... | https://github.com/huggingface/datasets/issues/4597 | [
"cc @huggingface/datasets: it seems like https://www.researchgate.net/ is flaky for datasets hosting (I put the \"hosted-on-google-drive\" tag since it's the same kind of issue I think)",
"Let's see if their license allows hosting their data on the Hub.",
"License is Creative Commons Attribution-NonCommercial-S... | null | 4,597 | false |
Dataset Viewer issue for universal_dependencies | ### Link
https://huggingface.co/datasets/universal_dependencies
### Description
invalid json response body at https://datasets-server.huggingface.co/splits?dataset=universal_dependencies reason: Unexpected token I in JSON at position 0
### Owner
_No response_ | https://github.com/huggingface/datasets/issues/4596 | [
"Thanks, looking at it!",
"Finally fixed! We updated the dataset viewer and it fixed the issue.\r\n\r\nhttps://huggingface.co/datasets/universal_dependencies/viewer/aqz_tudet/train\r\n\r\n<img width=\"1561\" alt=\"Capture d’écran 2022-09-07 à 13 29 18\" src=\"https://user-images.githubusercontent.com/1676121/18... | null | 4,596 | false |
Dataset Viewer issue with False positive PII redaction | ### Link
https://huggingface.co/datasets/cakiki/rosetta-code
### Description
Hello, I just noticed an entry being redacted that shouldn't have been:
`RootMeanSquare@Range[10]` is being displayed as `[email protected][10]`
### Owner
_No response_ | https://github.com/huggingface/datasets/issues/4595 | [
"The value is in the data, it's not an issue with the \"dataset-viewer\".\r\n\r\n<img width=\"1161\" alt=\"Capture d’écran 2022-06-29 à 10 25 51\" src=\"https://user-images.githubusercontent.com/1676121/176389325-4d2a9a7f-1583-45b8-aa7a-960ffaa6a36a.png\">\r\n\r\n Maybe open a PR: https://huggingface.co/datasets/... | null | 4,595 | false |
load_from_disk suggests incorrect fix when used to load DatasetDict | Edit: Please feel free to remove this issue. The problem was not the error message but the fact that the DatasetDict.load_from_disk does not support loading nested splits, i.e. if one of the splits is itself a DatasetDict. If nesting splits is an antipattern, perhaps the load_from_disk function can throw a warning indi... | https://github.com/huggingface/datasets/issues/4594 | [] | null | 4,594 | false |
Fix error message when using load_from_disk to load DatasetDict | Issue #4594
Issue: When `datasets.load_from_disk` is wrongly used to load a `DatasetDict`, the error message suggests using `datasets.load_from_disk`, which is the same function that generated the error.
Fix: The appropriate function which should be suggested instead is `datasets.dataset_dict.load_from_disk`.
Chan... | https://github.com/huggingface/datasets/pull/4593 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4593",
"html_url": "https://github.com/huggingface/datasets/pull/4593",
"diff_url": "https://github.com/huggingface/datasets/pull/4593.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4593.patch",
"merged_at": null
} | 4,593 | true |
Issue with jalFaizy/detect_chess_pieces when running datasets-cli test | ### Link
https://huggingface.co/datasets/jalFaizy/detect_chess_pieces
### Description
I am trying to write a appropriate data loader for [a custom dataset](https://huggingface.co/datasets/jalFaizy/detect_chess_pieces) using [this script](https://huggingface.co/datasets/jalFaizy/detect_chess_pieces/blob/main/detect_c... | https://github.com/huggingface/datasets/issues/4592 | [
"Hi @faizankshaikh\r\n\r\nPlease note that we have recently launched the Community feature, specifically targeted to create Discussions (about issues/questions/asking-for-help) on each Dataset on the Hub:\r\n- Blog post: https://huggingface.co/blog/community-update\r\n- Docs: https://huggingface.co/docs/hub/reposit... | null | 4,592 | false |
Can't push Images to hub with manual Dataset | ## Describe the bug
If I create a dataset including an 'Image' feature manually, when pushing to hub decoded images are not pushed,
instead it looks for image where image local path is/used to be.
This doesn't (at least didn't used to) happen with imagefolder. I want to build dataset manually because it is compli... | https://github.com/huggingface/datasets/issues/4591 | [
"Hi, thanks for reporting! This issue stems from the changes introduced in https://github.com/huggingface/datasets/pull/4282 (cc @lhoestq), in which list casts are ignored if they don't change the list type (required to preserve `null` values). And `push_to_hub` does a special cast to embed external image files but... | null | 4,591 | false |
Generalize meta_path json file creation in load.py [#4540] | # What does this PR do?
## Summary
*In function `_copy_script_and_other_resources_in_importable_dir`, using string split when generating `meta_path` throws error in the edge case raised in #4540.*
## Additions
-
## Changes
- Changed meta_path to use `os.path.splitext` instead of using `str.split` to gener... | https://github.com/huggingface/datasets/pull/4590 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@albertvillanova, Can you please review this PR for Issue #4540 ",
"@lhoestq Thank you for merging the PR . Is there a slack channel for contributing to the datasets library. I would love to work on the library and make meaningfu... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4590",
"html_url": "https://github.com/huggingface/datasets/pull/4590",
"diff_url": "https://github.com/huggingface/datasets/pull/4590.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4590.patch",
"merged_at": "2022-07-07T13:17... | 4,590 | true |
Permission denied: '/home/.cache' when load_dataset with local script | null | https://github.com/huggingface/datasets/issues/4589 | [] | null | 4,589 | false |
Host head_qa data on the Hub and fix NonMatchingChecksumError | This PR:
- Hosts head_qa data on the Hub instead of Google Drive
- Fixes NonMatchingChecksumError
Fix https://huggingface.co/datasets/head_qa/discussions/1 | https://github.com/huggingface/datasets/pull/4588 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"Hi @albertvillanova ! Thanks for the fix ;)\r\nCan I safely checkout from this branch to build `datasets` or it is preferable to wait until all CI tests pass?\r\nThanks 🙏 ",
"@younesbelkada we have just merged this PR."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4588",
"html_url": "https://github.com/huggingface/datasets/pull/4588",
"diff_url": "https://github.com/huggingface/datasets/pull/4588.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4588.patch",
"merged_at": "2022-07-05T15:49... | 4,588 | true |
Validate new_fingerprint passed by user | Users can pass the dataset fingerprint they want in `map` and other dataset transforms.
However the fingerprint is used to name cache files so we need to make sure it doesn't contain bad characters as mentioned in https://github.com/huggingface/datasets/issues/1718, and that it's not too long | https://github.com/huggingface/datasets/pull/4587 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4587",
"html_url": "https://github.com/huggingface/datasets/pull/4587",
"diff_url": "https://github.com/huggingface/datasets/pull/4587.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4587.patch",
"merged_at": "2022-06-28T14:00... | 4,587 | true |
Host pn_summary data on the Hub instead of Google Drive | Fix #4581. | https://github.com/huggingface/datasets/pull/4586 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4586",
"html_url": "https://github.com/huggingface/datasets/pull/4586",
"diff_url": "https://github.com/huggingface/datasets/pull/4586.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4586.patch",
"merged_at": "2022-06-28T14:42... | 4,586 | true |
Host multi_news data on the Hub instead of Google Drive | Host data files of multi_news dataset on the Hub.
They were on Google Drive.
Fix #4580. | https://github.com/huggingface/datasets/pull/4585 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4585",
"html_url": "https://github.com/huggingface/datasets/pull/4585",
"diff_url": "https://github.com/huggingface/datasets/pull/4585.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4585.patch",
"merged_at": "2022-06-28T14:08... | 4,585 | true |
Add binary classification task IDs | As a precursor to aligning the task IDs in `datasets` and AutoTrain, we need a way to distinguish binary vs multiclass vs multilabel classification.
This PR adds binary classification to the task IDs to enable this.
Related AutoTrain issue: https://github.com/huggingface/autonlp-backend/issues/597
cc @abhishek... | https://github.com/huggingface/datasets/pull/4584 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4584). All of your documentation changes will be reflected on that endpoint.",
"> Awesome thanks ! Can you add it to https://github.com/huggingface/hub-docs/blob/main/js/src/lib/interfaces/Types.ts first please ? This is where ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4584",
"html_url": "https://github.com/huggingface/datasets/pull/4584",
"diff_url": "https://github.com/huggingface/datasets/pull/4584.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4584.patch",
"merged_at": null
} | 4,584 | true |
<code> implementation of FLAC support using torchaudio | I had added Audio FLAC support with torchaudio given that Librosa and SoundFile can give problems. Also, FLAC is been used as audio from https://mlcommons.org/en/peoples-speech/ | https://github.com/huggingface/datasets/pull/4583 | [] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4583",
"html_url": "https://github.com/huggingface/datasets/pull/4583",
"diff_url": "https://github.com/huggingface/datasets/pull/4583.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4583.patch",
"merged_at": null
} | 4,583 | true |
add_column should preserve _indexes | https://github.com/huggingface/datasets/issues/3769#issuecomment-1167146126
doing `.add_column("x",x_data)` also removed any `_indexes` on the dataset, decided this shouldn't be the case.
This was because `add_column` was creating a new `Dataset(...)` and wasn't possible to pass indexes on init.
with this PR now... | https://github.com/huggingface/datasets/pull/4582 | [
"The docs for this PR live [here](https://moon-ci-docs.huggingface.co/docs/datasets/pr_4582). All of your documentation changes will be reflected on that endpoint."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4582",
"html_url": "https://github.com/huggingface/datasets/pull/4582",
"diff_url": "https://github.com/huggingface/datasets/pull/4582.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4582.patch",
"merged_at": null
} | 4,582 | true |
Dataset Viewer issue for pn_summary | ### Link
https://huggingface.co/datasets/pn_summary/viewer/1.0.0/validation
### Description
Getting an index error on the `validation` and `test` splits:
```
Server error
Status code: 400
Exception: IndexError
Message: list index out of range
```
### Owner
No | https://github.com/huggingface/datasets/issues/4581 | [
"linked to https://github.com/huggingface/datasets/issues/4580#issuecomment-1168373066?",
"Note that I refreshed twice this dataset, and I still have (another) error on one of the splits\r\n\r\n```\r\nStatus code: 400\r\nException: ClientResponseError\r\nMessage: 403, message='Forbidden', url=URL('htt... | null | 4,581 | false |
Dataset Viewer issue for multi_news | ### Link
https://huggingface.co/datasets/multi_news
### Description
Not sure what the index error is referring to here:
```
Status code: 400
Exception: IndexError
Message: list index out of range
```
### Owner
No | https://github.com/huggingface/datasets/issues/4580 | [
"Thanks for reporting, @lewtun.\r\n\r\nI forced the refreshing of the preview and it worked OK for train and validation splits.\r\n\r\nI guess the error has to do with the data files being hosted at Google Drive: this gives errors when requested automatically using scripts.\r\nWe should host them to fix the error. ... | null | 4,580 | false |
Support streaming cfq dataset | Support streaming cfq dataset. | https://github.com/huggingface/datasets/pull/4579 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"@lhoestq I've been refactoring a little the code:\r\n- Use less RAM by loading only the required samples: only if its index is in the splits file\r\n- Start yielding \"earlier\" in streaming mode: for each `split_idx`:\r\n - either ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4579",
"html_url": "https://github.com/huggingface/datasets/pull/4579",
"diff_url": "https://github.com/huggingface/datasets/pull/4579.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4579.patch",
"merged_at": "2022-07-04T19:23... | 4,579 | true |
[Multi Configs] Use directories to differentiate between subsets/configurations | Currently to define several subsets/configurations of your dataset, you need to use a dataset script.
However it would be nice to have a no-code way to to this.
For example we could specify different configurations of a dataset (for example, if a dataset contains different languages) with one directory per confi... | https://github.com/huggingface/datasets/issues/4578 | [
"I want to be able to create folders in a model.",
"How to set new split names, instead of train/test/validation? For example, I have a local dataset, consists of several subsets, named \"A\", \"B\", and \"C\". How can I create a huggingface dataset, with splits A/B/C ?\r\n\r\nThe document in https://huggingface.... | null | 4,578 | false |
Add authentication tip to `load_dataset` | Add an authentication tip similar to the one in transformers' `PreTrainedModel.from_pretrained` to `load_dataset`/`load_dataset_builder`. | https://github.com/huggingface/datasets/pull/4577 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4577",
"html_url": "https://github.com/huggingface/datasets/pull/4577",
"diff_url": "https://github.com/huggingface/datasets/pull/4577.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4577.patch",
"merged_at": "2022-07-04T13:01... | 4,577 | true |
Include `metadata.jsonl` in resolved data files | Include `metadata.jsonl` in resolved data files.
Fix #4548
@lhoestq ~~https://github.com/huggingface/datasets/commit/d94336d30eef17fc9abc67f67fa1c139661f4e75 adds support for metadata files placed at the root, and https://github.com/huggingface/datasets/commit/4d20618ea7a19bc143ddc5fdff9d79e671fcbb95 accounts fo... | https://github.com/huggingface/datasets/pull/4576 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"I still don't know if the way we implemented data files resolution could support the metadata.jsonl file without bad side effects for the other packaged builders. In particular here if you have a folder of csv/parquet/whatever files ... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4576",
"html_url": "https://github.com/huggingface/datasets/pull/4576",
"diff_url": "https://github.com/huggingface/datasets/pull/4576.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4576.patch",
"merged_at": "2022-06-30T10:15... | 4,576 | true |
Problem about wmt17 zh-en dataset | It seems that in subset casia2015, some samples are like `{'c[hn]':'xxx', 'en': 'aa'}`.
So when using `data = load_dataset('wmt17', "zh-en")` to load the wmt17 zh-en dataset, which will raise the exception:
```
Traceback (most recent call last):
File "train.py", line 78, in <module>
data = load_dataset(args.... | https://github.com/huggingface/datasets/issues/4575 | [
"Running into the same error with `wmt17/zh-en`, `wmt18/zh-en` and `wmt19/zh-en`.",
"@albertvillanova @lhoestq Could you take a look at this issue?",
"@winterfell2021 Hi, I wonder where the code you provided should be added. I tried to add them in the `datasets/table.py` in `array_cast` function, however, the '... | null | 4,575 | false |
Support streaming mlsum dataset | Support streaming mlsum dataset.
This PR:
- pins `fsspec` min version with fixed BlockSizeError: `fsspec[http]>=2021.11.1`
- https://github.com/fsspec/filesystem_spec/pull/830
- unpins `s3fs==2021.08.1` to align it with `fsspec` requirement: `s3fs>=2021.11.1`
> s3fs 2021.8.1 requires fsspec==2021.08.1
- s... | https://github.com/huggingface/datasets/pull/4574 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"After unpinning `s3fs` and pinning `fsspec[http]>=2021.11.1`, the CI installs\r\n- `fsspec-2022.1.0`\r\n- `s3fs-0.5.1`\r\n\r\nand raises the following error:\r\n```\r\n ImportError while loading conftest '/home/runner/work/datasets/d... | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4574",
"html_url": "https://github.com/huggingface/datasets/pull/4574",
"diff_url": "https://github.com/huggingface/datasets/pull/4574.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4574.patch",
"merged_at": "2022-07-21T12:40... | 4,574 | true |
Fix evaluation metadata for ncbi_disease | This PR fixes the task in the evaluation metadata and removes the metrics info as we've decided this is not a great way to propagate this information downstream. | https://github.com/huggingface/datasets/pull/4573 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4573",
"html_url": "https://github.com/huggingface/datasets/pull/4573",
"diff_url": "https://github.com/huggingface/datasets/pull/4573.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4573.patch",
"merged_at": null
} | 4,573 | true |
Dataset Viewer issue for mlsum | ### Link
https://huggingface.co/datasets/mlsum/viewer/de/train
### Description
There's seems to be a problem with the download / streaming of this dataset:
```
Server error
Status code: 400
Exception: BadZipFile
Message: File is not a zip file
```
### Owner
No | https://github.com/huggingface/datasets/issues/4572 | [
"Thanks for reporting, @lewtun.\r\n\r\nAfter investigation, it seems that the server https://gitlab.lip6.fr does not allow HTTP Range requests.\r\n\r\nWe are trying to find a workaround..."
] | null | 4,572 | false |
Dataset Viewer issue for gsarti/flores_101 | ### Link
https://huggingface.co/datasets/gsarti/flores_101
### Description
It seems like streaming isn't supported for this dataset:
```
Server Error
Status code: 400
Exception: NotImplementedError
Message: Extraction protocol for TAR archives like 'https://dl.fbaipublicfiles.com/flores101/dataset... | https://github.com/huggingface/datasets/issues/4571 | [
"Related to https://github.com/huggingface/datasets/issues/4562#issuecomment-1166911751\r\n\r\nI'll assign @albertvillanova ",
"I'm just wondering why we don't have this dataset under:\r\n- the `facebook` namespace\r\n- or the canonical dataset `flores`: why does this only have 2 languages?"
] | null | 4,571 | false |
Dataset sharding non-contiguous? | ## Describe the bug
I'm not sure if this is a bug; more likely normal behavior but i wanted to double check.
Is it normal that `datasets.shard` does not produce chunks that, when concatenated produce the original ordering of the sharded dataset?
This might be related to this pull request (https://github.com/huggi... | https://github.com/huggingface/datasets/issues/4570 | [
"This was silly; I was sure I'd looked for a `contiguous` argument, and was certain there wasn't one the first time I looked :smile:\r\n\r\nSorry about that.",
"Hi! You can pass `contiguous=True` to `.shard()` get contiguous shards. More info on this and the default behavior can be found in the [docs](https://hug... | null | 4,570 | false |
Dataset Viewer issue for sst2 | ### Link
https://huggingface.co/datasets/sst2
### Description
Not sure what is causing this, however it seems that `load_dataset("sst2")` also hangs (even though it downloads the files without problem):
```
Status code: 400
Exception: Exception
Message: Give up after 5 attempts with Connectio... | https://github.com/huggingface/datasets/issues/4569 | [
"Hi @lewtun, thanks for reporting.\r\n\r\nI have checked locally and refreshed the preview and it seems working smooth now:\r\n```python\r\nIn [8]: ds\r\nOut[8]: \r\nDatasetDict({\r\n train: Dataset({\r\n features: ['idx', 'sentence', 'label'],\r\n num_rows: 67349\r\n })\r\n validation: Datas... | null | 4,569 | false |
XNLI cache reload is very slow | ### Reproduce
Using `2.3.3.dev0`
`from datasets import load_dataset`
`load_dataset("xnli", "en")`
Turn off Internet
`load_dataset("xnli", "en")`
I cancelled the second `load_dataset` eventually cuz it took super long. It would be great to have something to specify e.g. `only_load_from_cache` and avoid the ... | https://github.com/huggingface/datasets/issues/4568 | [
"Hi,\r\nCould you tell us how you are running this code?\r\nI tested on my machine (M1 Mac). And it is running fine both on and off internet.\r\n\r\n<img width=\"1033\" alt=\"Screen Shot 2022-07-03 at 1 32 25 AM\" src=\"https://user-images.githubusercontent.com/8711912/177026364-4ad7cedb-e524-4513-97f7-7961bbb34c90... | null | 4,568 | false |
Add evaluation data for amazon_reviews_multi | null | https://github.com/huggingface/datasets/pull/4567 | [
"_The documentation is not available anymore as the PR was closed or merged._",
"As discussed with @lewtun, we are closing this PR, because it requires first the task names to be aligned between AutoTrain and datasets."
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4567",
"html_url": "https://github.com/huggingface/datasets/pull/4567",
"diff_url": "https://github.com/huggingface/datasets/pull/4567.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4567.patch",
"merged_at": null
} | 4,567 | true |
Document link #load_dataset_enhancing_performance points to nowhere | ## Describe the bug
A clear and concise description of what the bug is.

The [load_dataset_enhancing_performance](https://huggingface.co/docs/datasets/v2.3.2/en/package_reference/main_classes#load_dat... | https://github.com/huggingface/datasets/issues/4566 | [
"Hi! This is indeed the link the docstring should point to. Are you interested in submitting a PR to fix this?",
"https://github.com/huggingface/datasets/blame/master/docs/source/cache.mdx#L93\r\n\r\nThere seems already an anchor here. Somehow it doesn't work. I am not very familiar with how this online documenta... | null | 4,566 | false |
Add UFSC OCPap dataset | ## Adding a Dataset
- **Name:** UFSC OCPap: Papanicolaou Stained Oral Cytology Dataset (v4)
- **Description:** The UFSC OCPap dataset comprises 9,797 labeled images of 1200x1600 pixels acquired from 5 slides of cancer diagnosed and 3 healthy of oral brush samples, from distinct patients.
- **Paper:** https://dx.doi.... | https://github.com/huggingface/datasets/issues/4565 | [
"I will add this directly on the hub (same as #4486)—in https://huggingface.co/lapix"
] | null | 4,565 | false |
Support streaming bookcorpus dataset | Support streaming bookcorpus dataset. | https://github.com/huggingface/datasets/pull/4564 | [
"_The documentation is not available anymore as the PR was closed or merged._"
] | {
"url": "https://api.github.com/repos/huggingface/datasets/pulls/4564",
"html_url": "https://github.com/huggingface/datasets/pull/4564",
"diff_url": "https://github.com/huggingface/datasets/pull/4564.diff",
"patch_url": "https://github.com/huggingface/datasets/pull/4564.patch",
"merged_at": "2022-07-06T09:23... | 4,564 | true |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.