title
stringlengths
1
290
body
stringlengths
0
228k
html_url
stringlengths
46
51
comments
list
pull_request
dict
number
int64
1
5.59k
is_pull_request
bool
2 classes
Fix sem_eval_2018_task_1 download location
This changes the download location of sem_eval_2018_task_1 files to include the test set labels as discussed in https://github.com/huggingface/datasets/issues/2745#issuecomment-954588500_ with @lhoestq.
https://github.com/huggingface/datasets/pull/3549
[ "Hi ! Thanks for pushing this :)\r\n\r\nIt seems that you created this PR from an old version of `datasets` that didn't have the sem_eval_2018_task_1.py file.\r\n\r\nCan you try merging `master` into your branch ? Or re-create your PR from a branch that comes from a more recent version of `datasets` ?\r\n\r\nAnd so...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3549", "html_url": "https://github.com/huggingface/datasets/pull/3549", "diff_url": "https://github.com/huggingface/datasets/pull/3549.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3549.patch", "merged_at": null }
3,549
true
Specify the feature types of a dataset on the Hub without needing a dataset script
**Is your feature request related to a problem? Please describe.** Currently if I upload a CSV with paths to audio files, the column type is string instead of Audio. **Describe the solution you'd like** I'd like to be able to specify the types of the column, so that when loading the dataset I directly get the feat...
https://github.com/huggingface/datasets/issues/3548
[ "After looking into this, discovered that this is already supported if the `dataset_infos.json` file is configured correctly! Here is a working example: https://huggingface.co/datasets/abidlabs/test-audio-13\r\n\r\nThis should be probably be documented, though. " ]
null
3,548
false
Datasets created with `push_to_hub` can't be accessed in offline mode
## Describe the bug In offline mode, one can still access previously-cached datasets. This fails with datasets created with `push_to_hub`. ## Steps to reproduce the bug in Python: ``` import datasets mpwiki = datasets.load_dataset("teven/matched_passages_wikidata") ``` in bash: ``` export HF_DATASETS_OFFLIN...
https://github.com/huggingface/datasets/issues/3547
[ "Thanks for reporting. I think this can be fixed by improving the `CachedDatasetModuleFactory` and making it look into the `parquet` cache directory (datasets from push_to_hub are loaded with the parquet dataset builder). I'll look into it", "Hi, I'm having the same issue. Is there any update on this?", "We hav...
null
3,547
false
Remove print statements in datasets
This is a second time I'm removing print statements in our datasets, so I've added a test to avoid these issues in the future.
https://github.com/huggingface/datasets/pull/3546
[ "The CI failures are unrelated to the changes." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3546", "html_url": "https://github.com/huggingface/datasets/pull/3546", "diff_url": "https://github.com/huggingface/datasets/pull/3546.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3546.patch", "merged_at": "2022-01-07T18:09...
3,546
true
fix: 🐛 pass token when retrieving the split names
null
https://github.com/huggingface/datasets/pull/3545
[ "Currently, it does not work with https://huggingface.co/datasets/mozilla-foundation/common_voice_7_0/blob/main/common_voice_7_0.py#L146 (which was the goal), because `dl_manager.download_config.use_auth_token` is ignored, and the authentication is required to be use `huggingface-cli login`.\r\nIn my use case (data...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3545", "html_url": "https://github.com/huggingface/datasets/pull/3545", "diff_url": "https://github.com/huggingface/datasets/pull/3545.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3545.patch", "merged_at": "2022-01-10T10:51...
3,545
true
Ability to split a dataset in multiple files.
Hello, **Is your feature request related to a problem? Please describe.** My use case is that I have one writer that adds columns and multiple workers reading the same `Dataset`. Each worker should have access to columns added by the writer when they reload the dataset. I understand that we shouldn't overwrite...
https://github.com/huggingface/datasets/issues/3544
[]
null
3,544
false
Allow loading community metrics from the hub, just like datasets
**Is your feature request related to a problem? Please describe.** Currently, I can load a metric implemented by me by providing the local path to the file in `load_metric`. However, there is no option to do it with the metric uploaded to the hub. This means that if I want to allow other users to use it, they must d...
https://github.com/huggingface/datasets/issues/3543
[ "Hi ! Thanks for your message :) This is a great idea indeed. We haven't started working on this yet though. For now I guess you can host your metric on the Hub (either with your model or your dataset) and use `hf_hub_download` to download it (docs [here](https://github.com/huggingface/huggingface_hub/blob/main/doc...
null
3,543
false
Update the CC-100 dataset card
* summary from the dataset homepage * more details about the data structure * this dataset does not contain annotations
https://github.com/huggingface/datasets/pull/3542
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3542", "html_url": "https://github.com/huggingface/datasets/pull/3542", "diff_url": "https://github.com/huggingface/datasets/pull/3542.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3542.patch", "merged_at": "2022-01-06T18:37...
3,542
true
Support 7-zip compressed data files
**Is your feature request related to a problem? Please describe.** We should support 7-zip compressed data files: - [x] in `extract`: - #4672 - [ ] in `iter_archive`: for streaming mode both in streaming and non-streaming modes.
https://github.com/huggingface/datasets/issues/3541
[ "This should also resolve: https://github.com/huggingface/datasets/issues/3185." ]
null
3,541
false
How to convert torch.utils.data.Dataset to datasets.arrow_dataset.Dataset?
Hi, I use torch.utils.data.Dataset to define my own data, but I need to use the 'map' function of datasets.arrow_dataset.Dataset later, so I hope to convert torch.utils.data.Dataset to datasets.arrow_dataset.Dataset. Here is an example. ``` from torch.utils.data import Dataset from datasets.arrow_dataset import ...
https://github.com/huggingface/datasets/issues/3540
[]
null
3,540
false
Research wording for nc licenses
null
https://github.com/huggingface/datasets/pull/3539
[ "The CI failure is about some missing tags or sections in the dataset cards, and is unrelated to the part about non commercial use of this PR. Merging" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3539", "html_url": "https://github.com/huggingface/datasets/pull/3539", "diff_url": "https://github.com/huggingface/datasets/pull/3539.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3539.patch", "merged_at": "2022-01-06T18:58...
3,539
true
Readme usage update
Noticing that the recent commit throws a lot of errors in the automatic checks. It looks to me that those errors are simply errors that were already there (metadata issues), unrelated to what I've just changed, but worth another look to make sure.
https://github.com/huggingface/datasets/pull/3538
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3538", "html_url": "https://github.com/huggingface/datasets/pull/3538", "diff_url": "https://github.com/huggingface/datasets/pull/3538.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3538.patch", "merged_at": "2022-01-05T23:24...
3,538
true
added PII statements and license links to data cards
Updates for the following datacards: multilingual_librispeech openslr speech commands superb timit_asr vctk
https://github.com/huggingface/datasets/pull/3537
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3537", "html_url": "https://github.com/huggingface/datasets/pull/3537", "diff_url": "https://github.com/huggingface/datasets/pull/3537.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3537.patch", "merged_at": "2022-01-05T22:02...
3,537
true
update `pretty_name` for all datasets
This PR updates `pretty_name` for all datasets. Previous PR #3498 had done this for only first 200 datasets
https://github.com/huggingface/datasets/pull/3536
[ "Pushed the lastest changes!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3536", "html_url": "https://github.com/huggingface/datasets/pull/3536", "diff_url": "https://github.com/huggingface/datasets/pull/3536.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3536.patch", "merged_at": "2022-01-12T22:59...
3,536
true
Add SVHN dataset
Add the SVHN dataset. Additional notes: * compared to the TFDS implementation, exposes additional the "full numbers" config * adds the streaming support for `os.path.splitext` and `scipy.io.loadmat` * adds `h5py` to the requirements list for the dummy data test
https://github.com/huggingface/datasets/pull/3535
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3535", "html_url": "https://github.com/huggingface/datasets/pull/3535", "diff_url": "https://github.com/huggingface/datasets/pull/3535.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3535.patch", "merged_at": "2022-01-12T14:14...
3,535
true
Update wiki_dpr README.md
Some infos of wiki_dpr were missing as noted in https://github.com/huggingface/datasets/issues/3510, I added them and updated the tags and the examples Close #3510.
https://github.com/huggingface/datasets/pull/3534
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3534", "html_url": "https://github.com/huggingface/datasets/pull/3534", "diff_url": "https://github.com/huggingface/datasets/pull/3534.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3534.patch", "merged_at": "2022-01-05T14:16...
3,534
true
Task search function on hub not working correctly
When I want to look at all datasets of the category: `speech-processing` *i.e.* https://huggingface.co/datasets?task_categories=task_categories:speech-processing&sort=downloads , then the following dataset doesn't show up for some reason: - https://huggingface.co/datasets/speech_commands even thought it's task t...
https://github.com/huggingface/datasets/issues/3533
[ "known issue due to https://github.com/huggingface/datasets/pull/2362 (and [internal](https://github.com/huggingface/moon-landing/issues/946)) , will be solved soon", "hmm actually i have no recollection of why I said that", "Because it has dots in some YAML keys, it can't be parsed and indexed by the back-end"...
null
3,533
false
Give clearer instructions to add the YAML tags
Fix #3531. CC: @julien-c @VictorSanh
https://github.com/huggingface/datasets/pull/3532
[ "this is great, maybe just put all of it in one line?\r\n\r\n> TODO: Add YAML tags here. Copy-paste the tags obtained with the online tagging app: https://huggingface.co/spaces/huggingface/datasets-tagging" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3532", "html_url": "https://github.com/huggingface/datasets/pull/3532", "diff_url": "https://github.com/huggingface/datasets/pull/3532.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3532.patch", "merged_at": "2022-01-17T15:54...
3,532
true
Give clearer instructions to add the YAML tags
## Describe the bug As reported by @julien-c, many community datasets contain the line `YAML tags:` at the top of the YAML section in the header of the README file. See e.g.: https://huggingface.co/datasets/bigscience/P3/commit/a03bea08cf4d58f268b469593069af6aeb15de32 Maybe we should give clearer instruction/hints...
https://github.com/huggingface/datasets/issues/3531
[]
null
3,531
false
Update README.md
Removing reference to "Common Voice" in Personal and Sensitive Information section. Adding link to license. Correct license type in metadata.
https://github.com/huggingface/datasets/pull/3530
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3530", "html_url": "https://github.com/huggingface/datasets/pull/3530", "diff_url": "https://github.com/huggingface/datasets/pull/3530.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3530.patch", "merged_at": "2022-01-05T12:50...
3,530
true
Update README.md
Updating licensing information & personal and sensitive information.
https://github.com/huggingface/datasets/pull/3529
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3529", "html_url": "https://github.com/huggingface/datasets/pull/3529", "diff_url": "https://github.com/huggingface/datasets/pull/3529.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3529.patch", "merged_at": "2022-01-05T12:50...
3,529
true
Update README.md
Updating license with appropriate capitalization & a link. Updating Personal and Sensitive Information to address PII concern.
https://github.com/huggingface/datasets/pull/3528
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3528", "html_url": "https://github.com/huggingface/datasets/pull/3528", "diff_url": "https://github.com/huggingface/datasets/pull/3528.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3528.patch", "merged_at": "2022-01-05T12:49...
3,528
true
Update README.md
Adding licensing information.
https://github.com/huggingface/datasets/pull/3527
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3527", "html_url": "https://github.com/huggingface/datasets/pull/3527", "diff_url": "https://github.com/huggingface/datasets/pull/3527.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3527.patch", "merged_at": "2022-01-05T00:23...
3,527
true
Update license to bookcorpus dataset card
Not entirely sure, following the links here, but it seems the relevant license is at https://github.com/soskek/bookcorpus/blob/master/LICENSE
https://github.com/huggingface/datasets/pull/3526
[ "The smashwords ToS apply for this dataset, we did the same for https://github.com/huggingface/datasets/pull/3525", "_The documentation is not available anymore as the PR was closed or merged._" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3526", "html_url": "https://github.com/huggingface/datasets/pull/3526", "diff_url": "https://github.com/huggingface/datasets/pull/3526.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3526.patch", "merged_at": "2022-09-30T10:21...
3,526
true
Adding license information for Openbookcorpus
Not entirely sure, following the links here, but it seems the relevant license is at https://github.com/soskek/bookcorpus/blob/master/LICENSE
https://github.com/huggingface/datasets/pull/3525
[ "The MIT license seems to be for the crawling code, no ? Then maybe we can also redirect users to the [terms of smashwords.com](https://www.smashwords.com/about/tos) regarding copyrights, in particular the paragraph 10 for end-users. In particular it seems that end users can download and use the content \"for their...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3525", "html_url": "https://github.com/huggingface/datasets/pull/3525", "diff_url": "https://github.com/huggingface/datasets/pull/3525.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3525.patch", "merged_at": "2022-04-20T09:48...
3,525
true
Adding link to license.
null
https://github.com/huggingface/datasets/pull/3524
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3524", "html_url": "https://github.com/huggingface/datasets/pull/3524", "diff_url": "https://github.com/huggingface/datasets/pull/3524.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3524.patch", "merged_at": "2022-01-05T12:31...
3,524
true
Added links to licensing and PII message in vctk dataset
null
https://github.com/huggingface/datasets/pull/3523
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3523", "html_url": "https://github.com/huggingface/datasets/pull/3523", "diff_url": "https://github.com/huggingface/datasets/pull/3523.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3523.patch", "merged_at": "2022-01-06T19:33...
3,523
true
wmt19 is broken (zh-en)
## Describe the bug A clear and concise description of what the bug is. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset("wmt19", 'zh-en') ``` ## Expected results The dataset should download. ## Actual results `ConnectionError: Couldn't reach ftp://cwmt-wm...
https://github.com/huggingface/datasets/issues/3522
[ "This issue is not reproducible." ]
null
3,522
false
Vivos license update
Updated the license information with the link to the license text
https://github.com/huggingface/datasets/pull/3521
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3521", "html_url": "https://github.com/huggingface/datasets/pull/3521", "diff_url": "https://github.com/huggingface/datasets/pull/3521.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3521.patch", "merged_at": null }
3,521
true
Audio datacard update - first pass
Filling out data card "Personal and Sensitive Information" for speech datasets to note PII concerns
https://github.com/huggingface/datasets/pull/3520
[ "I'm not sure that we want to change the tags at the top of the cards by hand. Those are used to create the tags in the hub. Although looking at all the tags now, we might want to normalize the current tags again (hyphens or no, \".0\" or no). Maybe we could add a binary tag for public domain or not?", "> \r\n\r\...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3520", "html_url": "https://github.com/huggingface/datasets/pull/3520", "diff_url": "https://github.com/huggingface/datasets/pull/3520.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3520.patch", "merged_at": "2022-01-05T12:30...
3,520
true
CC100: Using HTTPS for the data source URL fixes load_dataset()
Without this change the following script (with any lang parameter) consistently fails. After changing to the HTTPS URL, the script works as expected. ```python from datasets import load_dataset dataset = load_dataset("cc100", lang="en") ``` This is the error produced by the previous script: ```sh Using cus...
https://github.com/huggingface/datasets/pull/3519
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3519", "html_url": "https://github.com/huggingface/datasets/pull/3519", "diff_url": "https://github.com/huggingface/datasets/pull/3519.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3519.patch", "merged_at": "2022-01-05T17:28...
3,519
true
Add PubMed Central Open Access dataset
## Adding a Dataset - **Name:** PubMed Central Open Access - **Description:** The PMC Open Access Subset includes more than 3.4 million journal articles and preprints that are made available under license terms that allow reuse. - **Paper:** *link to the dataset paper if available* - **Data:** https://www.ncbi.nlm....
https://github.com/huggingface/datasets/issues/3518
[ "In the framework of BigScience:\r\n- bigscience-workshop/data_tooling#121\r\n\r\nI have created this dataset as a community dataset: https://huggingface.co/datasets/albertvillanova/pmc_open_access\r\n\r\nHowever, I was wondering that it may be more appropriate to move it under an org namespace: `pubmed_central` or...
null
3,518
false
Add CPPE-5 dataset
Adds the recently released CPPE-5 dataset.
https://github.com/huggingface/datasets/pull/3517
[ "Thanks so much, @mariosasko and @lhoestq , much appreciated!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3517", "html_url": "https://github.com/huggingface/datasets/pull/3517", "diff_url": "https://github.com/huggingface/datasets/pull/3517.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3517.patch", "merged_at": "2022-01-05T18:53...
3,517
true
dataset `asset` - change to raw.githubusercontent.com URLs
Changed the URLs to the ones it was automatically re-directing. Before, the download was failing
https://github.com/huggingface/datasets/pull/3516
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3516", "html_url": "https://github.com/huggingface/datasets/pull/3516", "diff_url": "https://github.com/huggingface/datasets/pull/3516.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3516.patch", "merged_at": "2022-01-03T17:39...
3,516
true
`ExpectedMoreDownloadedFiles` for `evidence_infer_treatment`
## Describe the bug I am trying to load a dataset called `evidence_infer_treatment`. The first subset (`1.1`) works fine but the second returns an error (`2.0`). It downloads a file but crashes during the checksums. ## Steps to reproduce the bug ```python >>> from datasets import load_dataset >>> load_dataset("e...
https://github.com/huggingface/datasets/issues/3515
[ "Thanks for reporting @VictorSanh.\r\n\r\nI'm looking at it... " ]
null
3,515
false
Fix to_tf_dataset references in docs
Fix the `to_tf_dataset` references in the docs. The currently failing example of usage will be fixed by #3338.
https://github.com/huggingface/datasets/pull/3514
[ "The code snippet in [this section](https://huggingface.co/docs/datasets/master/use_dataset.html?highlight=to_tf_dataset#tensorflow) is missing an import (`DataCollatorWithPadding`) and doesn't initialize the TF model before the `model.fit` call." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3514", "html_url": "https://github.com/huggingface/datasets/pull/3514", "diff_url": "https://github.com/huggingface/datasets/pull/3514.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3514.patch", "merged_at": "2022-01-05T18:52...
3,514
true
Add desc parameter to filter
Fix #3317
https://github.com/huggingface/datasets/pull/3513
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3513", "html_url": "https://github.com/huggingface/datasets/pull/3513", "diff_url": "https://github.com/huggingface/datasets/pull/3513.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3513.patch", "merged_at": "2022-01-05T18:31...
3,513
true
No Data format found
## Dataset viewer issue for '*name of the dataset*' **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
https://github.com/huggingface/datasets/issues/3512
[ "Hi, which dataset is giving you an error?" ]
null
3,512
false
Dataset
## Dataset viewer issue for '*name of the dataset*' **Link:** *link to the dataset viewer page* *short description of the issue* Am I the one who added this dataset ? Yes-No
https://github.com/huggingface/datasets/issues/3511
[ "Can you reopen with the correct dataset name (if relevant)?\r\n\r\nThanks", "The dataset viewer was down tonight. It works again." ]
null
3,511
false
`wiki_dpr` details for Open Domain Question Answering tasks
Hey guys! Thanks for creating the `wiki_dpr` dataset! I am currently trying to use the dataset for context retrieval using DPR on NQ questions and need details about what each of the files and data instances mean, which version of the Wikipedia dump it uses, etc. Please respond at your earliest convenience regard...
https://github.com/huggingface/datasets/issues/3510
[ "Hi ! According to the DPR paper, the wikipedia dump is the one from Dec. 20, 2018.\r\nEach instance contains a paragraph of at most 100 word, as well as the title of the wikipedia page it comes from and the DPR embedding (a 768-d vector).", "Closed by:\r\n- #3534" ]
null
3,510
false
Discuss whether support canonical datasets w/o dataset_infos.json and/or dummy data
I open this PR to have a public discussion about this topic and make a decision. As previously discussed, once we have the metadata in the dataset card (README file, containing both Markdown info and YAML tags), what is the point of having also the JSON metadata (dataset_infos.json file)? On the other hand, the d...
https://github.com/huggingface/datasets/issues/3507
[ "IMO, the data streaming test is good enough of a test that the dataset works correctly (assuming that we can more or less ensure that if streaming works then the non-streaming case will also work), so that for datasets that have a working dataset preview, we can remove the dummy data IMO. On the other hand, it see...
null
3,507
false
Allows DatasetDict.filter to have batching option
- Related to: #3244 - Fixes: #3503 We extends `.filter( ... batched: bool)` support to DatasetDict.
https://github.com/huggingface/datasets/pull/3506
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3506", "html_url": "https://github.com/huggingface/datasets/pull/3506", "diff_url": "https://github.com/huggingface/datasets/pull/3506.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3506.patch", "merged_at": "2022-01-04T10:24...
3,506
true
cast_column function not working with map function in streaming mode for Audio features
## Describe the bug I am trying to use Audio class for loading audio features using custom dataset. I am able to cast 'audio' feature into 'Audio' format with cast_column function. On using map function, I am not getting 'Audio' casted feature but getting path of audio file only. I am getting features of 'audio' of s...
https://github.com/huggingface/datasets/issues/3505
[ "Hi! This is probably due to the fact that `IterableDataset.map` sets `features` to `None` before mapping examples. We can fix the issue by passing the old `features` dict to the map generator and performing encoding/decoding there (before calling the map transform function)." ]
null
3,505
false
Unable to download PUBMED_title_abstracts_2019_baseline.jsonl.zst
## Describe the bug I am unable to download the PubMed dataset from the link provided in the [Hugging Face Course (Chapter 5 Section 4)](https://huggingface.co/course/chapter5/4?fw=pt). https://the-eye.eu/public/AI/pile_preliminary_components/PUBMED_title_abstracts_2019_baseline.jsonl.zst ## Steps to reproduce ...
https://github.com/huggingface/datasets/issues/3504
[ "Hi @ToddMorrill, thanks for reporting.\r\n\r\nThree weeks ago I contacted the team who created the Pile dataset to report this issue with their data host server: https://the-eye.eu\r\n\r\nThey told me that unfortunately, the-eye was heavily affected by the recent tornado catastrophe in the US. They hope to have th...
null
3,504
false
Batched in filter throws error
I hope this is really a bug, I could not find it among the open issues ## Describe the bug using `batched=False` in DataSet.filter throws error ```python TypeError: filter() got an unexpected keyword argument 'batched' ``` but in the docs it is lister as an argument. ## Steps to reproduce the bug ```python ...
https://github.com/huggingface/datasets/issues/3503
[]
null
3,503
false
Add QuALITY
Fixes #3441.
https://github.com/huggingface/datasets/pull/3502
[ "Thanks for your contribution, @jaketae. Are you still interested in adding this dataset?\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you create this dataset there. Please, feel free to tell us if y...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3502", "html_url": "https://github.com/huggingface/datasets/pull/3502", "diff_url": "https://github.com/huggingface/datasets/pull/3502.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3502.patch", "merged_at": null }
3,502
true
Update pib dataset card
Related to #3496
https://github.com/huggingface/datasets/pull/3501
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3501", "html_url": "https://github.com/huggingface/datasets/pull/3501", "diff_url": "https://github.com/huggingface/datasets/pull/3501.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3501.patch", "merged_at": "2021-12-29T11:13...
3,501
true
Docs: Add VCTK dataset description
This PR is a very minor followup to #1837, with only docs changes (single comment string).
https://github.com/huggingface/datasets/pull/3500
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3500", "html_url": "https://github.com/huggingface/datasets/pull/3500", "diff_url": "https://github.com/huggingface/datasets/pull/3500.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3500.patch", "merged_at": "2022-01-04T10:25...
3,500
true
Adjusting chunk size for streaming datasets
**Is your feature request related to a problem? Please describe.** I want to use mc4 which I cannot save locally, so I stream it. However, I want to process the entire dataset and filter some documents from it. With the current chunk size of around 1000 documents (right?) I hit a performance bottleneck because of the ...
https://github.com/huggingface/datasets/issues/3499
[ "Hi ! Data streaming uses `fsspec` to read the data files progressively. IIRC the block size for buffering is 5MiB by default. So every time you finish iterating over a block, it downloads the next one. You can still try to increase the `fsspec` block size for buffering if it can help. To do so you just need to inc...
null
3,499
false
update `pretty_name` for first 200 datasets
I made a script some time back to fetch `pretty_names` from `papers_with_code` dataset along with some other rules incase that dataset wasn't available on `papers_with_code`. Updating them in the `README` of `datasets`. Took only the first 200 datasets into consideration and after some eyeballing, most of them were loo...
https://github.com/huggingface/datasets/pull/3498
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3498", "html_url": "https://github.com/huggingface/datasets/pull/3498", "diff_url": "https://github.com/huggingface/datasets/pull/3498.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3498.patch", "merged_at": "2022-01-05T16:38...
3,498
true
Changing sampling rate in audio dataset and subsequently mapping with `num_proc > 1` leads to weird bug
Running: ```python from datasets import load_dataset, DatasetDict import datasets from transformers import AutoFeatureExtractor raw_datasets = DatasetDict() raw_datasets["train"] = load_dataset("common_voice", "ab", split="train") feature_extractor = AutoFeatureExtractor.from_pretrained("facebook/wav2ve...
https://github.com/huggingface/datasets/issues/3497
[ "Same error occures when using max samples with https://github.com/huggingface/transformers/blob/master/examples/pytorch/speech-recognition/run_speech_recognition_seq2seq.py", "I'm seeing this too, when using preprocessing_num_workers with \r\nhttps://github.com/huggingface/transformers/blob/master/examples/pytor...
null
3,497
false
Update version of pib dataset and make it streamable
This PR: - Updates version of pib dataset: from 0.0.0 to 1.3.0 - Makes the dataset streamable Fix #3491. CC: @severo
https://github.com/huggingface/datasets/pull/3496
[ "It seems like there is still an error: `Message: 'TarContainedFile' object has no attribute 'readable'`\r\n\r\nhttps://huggingface.co/datasets/pib/viewer", "@severo I was wondering about that...\r\n\r\nIt works fine when I run it in streaming mode in my terminal:\r\n```python\r\nIn [3]: from datasets impor...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3496", "html_url": "https://github.com/huggingface/datasets/pull/3496", "diff_url": "https://github.com/huggingface/datasets/pull/3496.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3496.patch", "merged_at": "2021-12-29T08:42...
3,496
true
Add VoxLingua107
## Adding a Dataset - **Name:** VoxLingua107 - **Description:** VoxLingua107 is a speech dataset for training spoken language identification models. - **Paper:** https://arxiv.org/abs/2011.12998 - **Data:** http://bark.phon.ioc.ee/voxlingua107/ - **Motivation:** 107 languages, totaling 6628 hours for the train sp...
https://github.com/huggingface/datasets/issues/3495
[]
null
3,495
false
Clone full repo to detect new tags when mirroring datasets on the Hub
The new releases of `datasets` were not detected because the shallow clone in the CI wasn't getting the git tags. By cloning the full repository we can properly detect a new release, and tag all the dataset repositories accordingly cc @SBrandeis
https://github.com/huggingface/datasets/pull/3494
[ "Good catch !!", "The CI fail is unrelated to this PR and fixed on master, merging :)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3494", "html_url": "https://github.com/huggingface/datasets/pull/3494", "diff_url": "https://github.com/huggingface/datasets/pull/3494.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3494.patch", "merged_at": "2021-12-28T16:07...
3,494
true
Fix VCTK encoding
utf-8 encoding was missing in the VCTK dataset builder added in #3351
https://github.com/huggingface/datasets/pull/3493
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3493", "html_url": "https://github.com/huggingface/datasets/pull/3493", "diff_url": "https://github.com/huggingface/datasets/pull/3493.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3493.patch", "merged_at": "2021-12-28T15:48...
3,493
true
Add `gzip` for `to_json`
(Partially) closes #3480. I have added `gzip` compression for `to_json`. I realised we can run into this compression problem with `to_csv` as well. `IOHandler` can be used for `to_csv` too. Please let me know if any changes are required.
https://github.com/huggingface/datasets/pull/3492
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3492", "html_url": "https://github.com/huggingface/datasets/pull/3492", "diff_url": "https://github.com/huggingface/datasets/pull/3492.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3492.patch", "merged_at": "2022-01-05T13:03...
3,492
true
Update version of pib dataset
On the Hub we have v0, while there exists v1.3. Related to bigscience-workshop/data_tooling#130
https://github.com/huggingface/datasets/issues/3491
[]
null
3,491
false
Does datasets support load text from HDFS?
The raw text data is stored on HDFS due to the dataset's size is too large to store on my develop machine, so I wander does datasets support read data from hdfs?
https://github.com/huggingface/datasets/issues/3490
[ "Hi ! `datasets` currently supports reading local files or files over HTTP. We may add support for other filesystems (cloud storages, hdfs...) at one point though :)" ]
null
3,490
false
Avoid unnecessary list creations
Like in `join([... for s in ...])`. Also changed other things that I saw: * Use a `with` statement for many `open` that missed them, so the files don't remain open. * Remove unused variables. * Many HTTP links converted into HTTPS (verified). * Remove unnecessary "r" mode arg in `open` (double-checked it was actual...
https://github.com/huggingface/datasets/pull/3489
[ "@bryant1410 Thanks for working on this. Could you please split the PR into 4 or 5 smaller PRs (ideally one PR for each bullet point from your description) because it's not practical to review such a large PR, especially if the changes are not interrelated?" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3489", "html_url": "https://github.com/huggingface/datasets/pull/3489", "diff_url": "https://github.com/huggingface/datasets/pull/3489.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3489.patch", "merged_at": null }
3,489
true
URL query parameters are set as path in the compression hop for fsspec
## Describe the bug There is an ssue with `StreamingDownloadManager._extract`. I don't know how the test `test_streaming_gg_drive_gzipped` passes: For ```python TEST_GG_DRIVE_GZIPPED_URL = "https://drive.google.com/uc?export=download&id=1Bt4Garpf0QLiwkJhHJzXaVa0I0H5Qhwz" urlpath = StreamingDownloadManager()....
https://github.com/huggingface/datasets/issues/3488
[ "I think the test passes because it simply ignore what's after `gzip://`.\r\n\r\nThe returned urlpath is expected to look like `gzip://filename::url`, and the filename is currently considered to be what's after the final `/`, hence the result.\r\n\r\nWe can decide to change this and simply have `gzip://::url`, this...
null
3,488
false
Update ADD_NEW_DATASET.md
fixed make style prompt for Windows Terminal
https://github.com/huggingface/datasets/pull/3487
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3487", "html_url": "https://github.com/huggingface/datasets/pull/3487", "diff_url": "https://github.com/huggingface/datasets/pull/3487.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3487.patch", "merged_at": "2021-12-27T15:00...
3,487
true
Fix weird spacing in ManualDownloadError message
`textwrap.dedent` works based on the spaces at the beginning. Before this change, there wasn't any space.
https://github.com/huggingface/datasets/pull/3486
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3486", "html_url": "https://github.com/huggingface/datasets/pull/3486", "diff_url": "https://github.com/huggingface/datasets/pull/3486.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3486.patch", "merged_at": "2021-12-28T09:00...
3,486
true
skip columns which cannot set to specific format when set_format
**Is your feature request related to a problem? Please describe.** When using `dataset.set_format("torch")`, I must make sure every columns in datasets can convert to `torch`, however, sometimes I want to keep some string columns. **Describe the solution you'd like** skip columns which cannot set to specific forma...
https://github.com/huggingface/datasets/issues/3485
[ "You can add columns that you wish to set into `torch` format using `dataset.set_format(\"torch\", ['id', 'abc'])` so that input batch of the transform only contains those columns", "Sorry, I miss `output_all_columns` args and thought after `dataset.set_format(\"torch\", columns=columns)` I can only get specific ...
null
3,485
false
make shape verification to use ArrayXD instead of nested lists for map
As describe in https://github.com/huggingface/datasets/issues/2005#issuecomment-793716753 and mentioned by @mariosasko in [image feature example](https://colab.research.google.com/drive/1mIrTnqTVkWLJWoBzT1ABSe-LFelIep1c#scrollTo=ow3XHDvf2I0B&line=1&uniqifier=1), IMO make shape verifcaiton to use ArrayXD instead of nest...
https://github.com/huggingface/datasets/issues/3484
[ "Hi! \r\n\r\nYes, this makes sense for numeric values, but first I have to finish https://github.com/huggingface/datasets/pull/3336 because currently ArrayXD only allows the first dimension to be dynamic." ]
null
3,484
false
Remove unused phony rule from Makefile
null
https://github.com/huggingface/datasets/pull/3483
[ "The CI failure is unrelated to this PR and fixed on master, merging !" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3483", "html_url": "https://github.com/huggingface/datasets/pull/3483", "diff_url": "https://github.com/huggingface/datasets/pull/3483.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3483.patch", "merged_at": "2022-01-05T16:34...
3,483
true
Fix duplicate keys in NewsQA
* Fix duplicate keys in NewsQA when loading from CSV files. * Fix s/narqa/newsqa/ in the download manually error message. * Make the download manually error message show nicely when printed. Otherwise, is hard to read due to spacing issues. * Fix the format of the license text. * Reformat the code to make it simple...
https://github.com/huggingface/datasets/pull/3482
[ "Flaky tests?", "Thanks for your contribution, @bryant1410.\r\n\r\nI think the fix of the duplicate key in this PR was superseded by:\r\n- #3696\r\n\r\nI'm closing this because we are moving all dataset scripts from GitHub to the Hugging Face Hub." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3482", "html_url": "https://github.com/huggingface/datasets/pull/3482", "diff_url": "https://github.com/huggingface/datasets/pull/3482.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3482.patch", "merged_at": null }
3,482
true
Fix overriding of filesystem info
Previously, `BaseCompressedFileFileSystem.info` was overridden and transformed from function to dict. This generated a bug for filesystem methods that use `self.info()`, like e.g. `fs.isfile()`. This PR: - Adds tests for `fs.isfile` (that use `fs.info`). - Fixes custom `BaseCompressedFileFileSystem.info` by rem...
https://github.com/huggingface/datasets/pull/3481
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3481", "html_url": "https://github.com/huggingface/datasets/pull/3481", "diff_url": "https://github.com/huggingface/datasets/pull/3481.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3481.patch", "merged_at": "2021-12-24T11:08...
3,481
true
the compression format requested when saving a dataset in json format is not respected
## Describe the bug In the documentation of the `to_json` method, it is stated in the parameters that > **to_json_kwargs – Parameters to pass to pandas’s pandas.DataFrame.to_json. however when we pass for example `compression="gzip"`, the saved file is not compressed. Would you also have expected compression t...
https://github.com/huggingface/datasets/issues/3480
[ "Thanks for reporting @SaulLu.\r\n\r\nAt first sight I think the problem is caused because `pandas` only takes into account the `compression` parameter if called with a non-null file path or buffer. And in our implementation, we call pandas `to_json` with `None` `path_or_buf`.\r\n\r\nWe should fix this:\r\n- either...
null
3,480
false
Dataset preview is not available (I think for all Hugging Face datasets)
## Dataset viewer issue for '*french_book_reviews*' **Link:** https://huggingface.co/datasets/Abirate/french_book_reviews **short description of the issue** For my dataset, the dataset preview is no longer functional (it used to work: The dataset had been added the day before and it was fine...) And, after lo...
https://github.com/huggingface/datasets/issues/3479
[ "You're right, we have an issue today with the datasets preview. We're investigating.", "It should be fixed now. Thanks for reporting.", "Down again. ", "Fixed for good." ]
null
3,479
false
Extend support for streaming datasets that use os.walk
This PR extends the support in streaming mode for datasets that use `os.walk`, by patching that function. This PR adds support for streaming mode to datasets: 1. autshumato 1. code_x_glue_cd_code_to_text 1. code_x_glue_tc_nl_code_search_adv 1. nchlt CC: @severo
https://github.com/huggingface/datasets/pull/3478
[ "Nice. I'll update the dataset viewer once merged, and test on these four datasets" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3478", "html_url": "https://github.com/huggingface/datasets/pull/3478", "diff_url": "https://github.com/huggingface/datasets/pull/3478.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3478.patch", "merged_at": "2021-12-24T10:50...
3,478
true
Use `iter_files` instead of `str(Path(...)` in image dataset
Use `iter_files` in the `beans` and the `cats_vs_dogs` dataset scripts as suggested by @albertvillanova. Additional changes: * Fix `iter_files` in `MockDownloadManager` (see this [CI error](https://app.circleci.com/pipelines/github/huggingface/datasets/9247/workflows/2657ff8a-b531-4fd9-a9fc-6541a72e8d83/jobs/57028)...
https://github.com/huggingface/datasets/pull/3477
[ "`iter_archive` is about to support ZIP archives. I think we should use this no ?\r\n\r\nsee #3347 https://github.com/huggingface/datasets/pull/3379", "I was interested in the support for isfile/dir in remote.\r\n\r\nAnyway, `iter_files` will be available for community users.", "I'm not a big fan of having two ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3477", "html_url": "https://github.com/huggingface/datasets/pull/3477", "diff_url": "https://github.com/huggingface/datasets/pull/3477.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3477.patch", "merged_at": "2021-12-28T15:15...
3,477
true
Extend support for streaming datasets that use ET.parse
This PR extends the support in streaming mode for datasets that use `ET.parse`, by patching the function. This PR adds support for streaming mode to datasets: 1. ami 1. assin 1. assin2 1. counter 1. enriched_web_nlg 1. europarl_bilingual 1. hyperpartisan_news_detection 1. polsum 1. qa4mre 1. quail 1. ted_...
https://github.com/huggingface/datasets/pull/3476
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3476", "html_url": "https://github.com/huggingface/datasets/pull/3476", "diff_url": "https://github.com/huggingface/datasets/pull/3476.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3476.patch", "merged_at": "2021-12-23T15:34...
3,476
true
The rotten_tomatoes dataset of movie reviews contains some reviews in Spanish
## Describe the bug See title. I don't think this is intentional and they probably should be removed. If they stay the dataset description should be at least updated to make it clear to the user. ## Steps to reproduce the bug Go to the [dataset viewer](https://huggingface.co/datasets/viewer/?dataset=rotten_tomato...
https://github.com/huggingface/datasets/issues/3475
[ "Hi @puzzler10, thanks for reporting.\r\n\r\nPlease note this dataset is not hosted on Hugging Face Hub. See: \r\nhttps://github.com/huggingface/datasets/blob/c8f914473b041833fd47178fa4373cdcb56ac522/datasets/rotten_tomatoes/rotten_tomatoes.py#L42\r\n\r\nIf there are issues with the source data of a dataset, you sh...
null
3,475
false
Decode images when iterating
If I iterate over a vision dataset, the images are not decoded, and the dictionary with the bytes is returned. This PR enables image decoding in `Dataset.__iter__` Close https://github.com/huggingface/datasets/issues/3473
https://github.com/huggingface/datasets/pull/3474
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3474", "html_url": "https://github.com/huggingface/datasets/pull/3474", "diff_url": "https://github.com/huggingface/datasets/pull/3474.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3474.patch", "merged_at": null }
3,474
true
Iterating over a vision dataset doesn't decode the images
## Describe the bug If I load `mnist` and I iterate over the dataset, the images are not decoded, and the dictionary with the bytes is returned. ## Steps to reproduce the bug ```python from datasets import load_dataset import PIL mnist = load_dataset("mnist", split="train") first_image = mnist[0]["image"...
https://github.com/huggingface/datasets/issues/3473
[ "As discussed, I remember I set `decoded=False` here to avoid decoding just by iterating over examples of dataset. We wanted to decode only if the \"audio\" field (for Audio feature) was accessed.", "> I set decoded=False here to avoid decoding just by iterating over examples of dataset. We wanted to decode only ...
null
3,473
false
Fix `str(Path(...))` conversion in streaming on Linux
Fix `str(Path(...))` conversion in streaming on Linux. This should fix the streaming of the `beans` and `cats_vs_dogs` datasets.
https://github.com/huggingface/datasets/pull/3472
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3472", "html_url": "https://github.com/huggingface/datasets/pull/3472", "diff_url": "https://github.com/huggingface/datasets/pull/3472.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3472.patch", "merged_at": "2021-12-22T16:52...
3,472
true
Fix Tashkeela dataset to yield stripped text
This PR: - Yields stripped text - Fix path for Windows - Adds license - Adds more info in dataset card Close bigscience-workshop/data_tooling#279
https://github.com/huggingface/datasets/pull/3471
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3471", "html_url": "https://github.com/huggingface/datasets/pull/3471", "diff_url": "https://github.com/huggingface/datasets/pull/3471.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3471.patch", "merged_at": "2021-12-22T10:12...
3,471
true
Fix rendering of docs
Minor fix in docs. Currently, `ClassLabel` docstring rendering is not right.
https://github.com/huggingface/datasets/pull/3470
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3470", "html_url": "https://github.com/huggingface/datasets/pull/3470", "diff_url": "https://github.com/huggingface/datasets/pull/3470.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3470.patch", "merged_at": "2021-12-22T09:23...
3,470
true
Fix METEOR missing NLTK's omw-1.4
NLTK 3.6.6 now requires `omw-1.4` to be downloaded for METEOR to work. This should fix the CI on master
https://github.com/huggingface/datasets/pull/3469
[ "I also modified the doctest call to raise the exception that doctest may catch, instead of `doctest.UnexpectedException`.\r\nThis will make debugging easier if it happens again" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3469", "html_url": "https://github.com/huggingface/datasets/pull/3469", "diff_url": "https://github.com/huggingface/datasets/pull/3469.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3469.patch", "merged_at": "2021-12-21T14:49...
3,469
true
Add COCO dataset
This PR adds the MS COCO dataset. Compared to the [TFDS](https://github.com/tensorflow/datasets/blob/master/tensorflow_datasets/object_detection/coco.py) script, this implementation adds 8 additional configs to cover the tasks other than object detection. Some notes: * the data exposed by TFDS is contained in the `...
https://github.com/huggingface/datasets/pull/3468
[ "The CI failures other than a missing dummy data file and missing fields in the card are unrelated to this PR. ", "Thanks a lot for this great work and fixing TFDS based script @mariosasko 🤗 will generate the dummy dataset and write the model card tomorrow!", "@mariosasko I added the dataset card, I'm on the d...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3468", "html_url": "https://github.com/huggingface/datasets/pull/3468", "diff_url": "https://github.com/huggingface/datasets/pull/3468.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3468.patch", "merged_at": null }
3,468
true
Push dataset infos.json to Hub
When doing `push_to_hub`, the feature types are lost (see issue https://github.com/huggingface/datasets/issues/3394). This PR fixes this by also pushing a `dataset_infos.json` file to the Hub, that stores the feature types. Other minor changes: - renamed the `___` separator to `--`, since `--` is now disallowed in...
https://github.com/huggingface/datasets/pull/3467
[ "The change from `___` to `--` was allowed by https://github.com/huggingface/moon-landing/pull/1657" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3467", "html_url": "https://github.com/huggingface/datasets/pull/3467", "diff_url": "https://github.com/huggingface/datasets/pull/3467.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3467.patch", "merged_at": "2021-12-21T17:00...
3,467
true
Add CRASS dataset
Added crass dataset
https://github.com/huggingface/datasets/pull/3466
[ "Hi Albert,\r\nThank you for your comments.\r\nI hope I have uploaded my local git repo to include the dummy files and style reworkings.\r\nAdded YAML in Readme as well.\r\n\r\nPlease check again.\r\n\r\nHope it works now :)", "Thanks for your contribution, @apergo-ai. \r\n\r\nWe are removing the dataset scripts ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3466", "html_url": "https://github.com/huggingface/datasets/pull/3466", "diff_url": "https://github.com/huggingface/datasets/pull/3466.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3466.patch", "merged_at": null }
3,466
true
Unable to load 'cnn_dailymail' dataset
## Describe the bug I wanted to load cnn_dailymail dataset from huggingface datasets on Google Colab, but I am getting an error while loading it. ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('cnn_dailymail', '3.0.0', ignore_verifications = True) ``` ## Expe...
https://github.com/huggingface/datasets/issues/3465
[ "Hi @talha1503, thanks for reporting.\r\n\r\nIt seems there is an issue with one of the data files hosted at Google Drive:\r\n```\r\nGoogle Drive - Quota exceeded\r\n\r\nSorry, you can't view or download this file at this time.\r\n\r\nToo many users have viewed or downloaded this file recently. Please try accessing...
null
3,465
false
struct.error: 'i' format requires -2147483648 <= number <= 2147483647
## Describe the bug A clear and concise description of what the bug is. using latest datasets=datasets-1.16.1-py3-none-any.whl process my own multilingual dataset by following codes, and the number of rows in all dataset is 306000, the max_length of each sentence is 256: ![image](https://user-images.githubusercon...
https://github.com/huggingface/datasets/issues/3464
[ "Hi ! Can you try setting `datasets.config.MAX_TABLE_NBYTES_FOR_PICKLING` to a smaller value than `4 << 30` (4GiB), for example `500 << 20` (500MiB) ? It should reduce the maximum size of the arrow table being pickled during multiprocessing.\r\n\r\nIf it fixes the issue, we can consider lowering the default value f...
null
3,464
false
Update swahili_news dataset
Update dataset with latest verion data files. Fix #3462. Close bigscience-workshop/data_tooling#107
https://github.com/huggingface/datasets/pull/3463
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3463", "html_url": "https://github.com/huggingface/datasets/pull/3463", "diff_url": "https://github.com/huggingface/datasets/pull/3463.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3463.patch", "merged_at": "2021-12-21T06:24...
3,463
true
Update swahili_news dataset
Please note also: the HuggingFace version at https://huggingface.co/datasets/swahili_news is outdated. An updated version, with deduplicated text and official splits, can be found at https://zenodo.org/record/5514203. ## Adding a Dataset - **Name:** swahili_news Instructions to add a new dataset can be found [he...
https://github.com/huggingface/datasets/issues/3462
[]
null
3,462
false
Fix links in metrics description
Remove Markdown syntax for links in metrics description, as it is not properly rendered. Related to #3437.
https://github.com/huggingface/datasets/pull/3461
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3461", "html_url": "https://github.com/huggingface/datasets/pull/3461", "diff_url": "https://github.com/huggingface/datasets/pull/3461.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3461.patch", "merged_at": "2021-12-20T17:14...
3,461
true
Don't encode lists as strings when using `Value("string")`
Following https://github.com/huggingface/datasets/pull/3456#event-5792250497 it looks like `datasets` can silently convert lists to strings using `str()`, instead of raising an error. This PR fixes this and should fix the issue with WER showing low values if the input format is not right.
https://github.com/huggingface/datasets/pull/3460
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3460", "html_url": "https://github.com/huggingface/datasets/pull/3460", "diff_url": "https://github.com/huggingface/datasets/pull/3460.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3460.patch", "merged_at": null }
3,460
true
dataset.filter overwriting previously set dataset._indices values, resulting in the wrong elements being selected.
## Describe the bug When using dataset.select to select a subset of a dataset, dataset._indices are set to indicate which elements are now considered in the dataset. The same thing happens when you shuffle the dataset; dataset._indices are set to indicate what the new order of the data is. However, if you then use a...
https://github.com/huggingface/datasets/issues/3459
[ "I think this is a duplicate of [#3190](https://github.com/huggingface/datasets/issues/3190)?", "Upgrading the datasets version as per #3190 fixes this bug. \r\nI'm Marking as closed." ]
null
3,459
false
Fix duplicated tag in wikicorpus dataset card
null
https://github.com/huggingface/datasets/pull/3458
[ "CI is failing just because of empty sections - merging" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3458", "html_url": "https://github.com/huggingface/datasets/pull/3458", "diff_url": "https://github.com/huggingface/datasets/pull/3458.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3458.patch", "merged_at": "2021-12-20T16:03...
3,458
true
Add CMU Graphics Lab Motion Capture dataset
## Adding a Dataset - **Name:** CMU Graphics Lab Motion Capture database - **Description:** The database contains free motions which you can download and use. - **Data:** http://mocap.cs.cmu.edu/ - **Motivation:** Nice motion capture dataset Instructions to add a new dataset can be found [here](https://github.c...
https://github.com/huggingface/datasets/issues/3457
[ "This dataset has files in ASF/AMC format. [ The skeleton file is the ASF file (Acclaim Skeleton File). The motion file is the AMC file (Acclaim Motion Capture data). ] \r\n\r\nSome questions : \r\n1. How do we go about representing these features using datasets.Features and generate examples ?\r\n2. The dataset d...
null
3,457
false
[WER] Better error message for wer
Currently we have the following problem when using the WER. When the input format to the WER metric is wrong, instead of throwing an error message a word-error-rate is computed which is incorrect. E.g. when doing the following: ```python from datasets import load_metric wer = load_metric("wer") target_str ...
https://github.com/huggingface/datasets/pull/3456
[ "Hi ! I don't think this would solve this issue.\r\nCurrently it looks like there's a bug that converts the list `[\"hello it's nice\"]` to a string `'[\"hello it's nice\"]'` since this is what the metric expects as input. The conversion is done before the data are passed to `_compute()`.\r\n\r\nThis is `Value(\"st...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3456", "html_url": "https://github.com/huggingface/datasets/pull/3456", "diff_url": "https://github.com/huggingface/datasets/pull/3456.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3456.patch", "merged_at": null }
3,456
true
Easier information editing
**Is your feature request related to a problem? Please describe.** It requires a lot of effort to improve a datasheet. **Describe the solution you'd like** UI or at least a link to the place where the code that needs to be edited is (and an easy way to edit this code directly from the site, without cloning, branc...
https://github.com/huggingface/datasets/issues/3455
[ "Hi ! I guess you are talking about the dataset cards that are in this repository on github ?\r\n\r\nI think github allows to submit a PR even for 1 line though the `Edit file` button on the page of the dataset card.\r\n\r\nMaybe let's mention this in `CONTRIBUTING.md` ?" ]
null
3,455
false
Fix iter_archive generator
This PR: - Adds tests to DownloadManager and StreamingDownloadManager `iter_archive` for both path and file inputs - Fixes bugs in `iter_archive` introduced in: - #3443 Fix #3453.
https://github.com/huggingface/datasets/pull/3454
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3454", "html_url": "https://github.com/huggingface/datasets/pull/3454", "diff_url": "https://github.com/huggingface/datasets/pull/3454.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3454.patch", "merged_at": "2021-12-20T10:04...
3,454
true
ValueError while iter_archive
## Describe the bug After the merge of: - #3443 the method `iter_archive` throws a ValueError: ``` ValueError: read of closed file ``` ## Steps to reproduce the bug ```python for path, file in dl_manager.iter_archive(archive_path): pass ```
https://github.com/huggingface/datasets/issues/3453
[]
null
3,453
false
why the stratify option is omitted from test_train_split function?
why the stratify option is omitted from test_train_split function? is there any other way implement the stratify option while splitting the dataset? as it is important point to be considered while splitting the dataset.
https://github.com/huggingface/datasets/issues/3452
[ "Hi ! It's simply not added yet :)\r\n\r\nIf someone wants to contribute to add the `stratify` parameter I'd be happy to give some pointers.\r\n\r\nIn the meantime, I guess you can use `sklearn` or other tools to do a stratified train/test split over the **indices** of your dataset and then do\r\n```\r\ntrain_datas...
null
3,452
false
[Staging] Update dataset repos automatically on the Hub
Let's have a script that updates the dataset repositories on staging for now. This way we can make sure it works fine before going in prod. Related to https://github.com/huggingface/datasets/issues/3341 The script runs on each commit on `master`. It checks the datasets that were changed, and it pushes the changes...
https://github.com/huggingface/datasets/pull/3451
[ "do keep us updated on how it's going in staging! cc @SBrandeis ", "Sure ! For now it works smoothly. We'll also do a new release today.\r\n\r\nI can send you some repos to explore on staging, in case you want to see how they look like after being updated.\r\nFor example [swahili_news](https://moon-staging.huggin...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3451", "html_url": "https://github.com/huggingface/datasets/pull/3451", "diff_url": "https://github.com/huggingface/datasets/pull/3451.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3451.patch", "merged_at": "2021-12-20T14:09...
3,451
true
Unexpected behavior doing Split + Filter
## Describe the bug I observed unexpected behavior when applying 'train_test_split' followed by 'filter' on dataset. Elements of the training dataset eventually end up in the test dataset (after applying the 'filter') ## Steps to reproduce the bug ``` from datasets import Dataset import pandas as pd dic = {'x'...
https://github.com/huggingface/datasets/issues/3450
[ "Hi ! This is an issue with `datasets` 1.12. Sorry for the inconvenience. Can you update to `>=1.13` ?\r\nsee https://github.com/huggingface/datasets/issues/3190\r\n\r\nMaybe we should also backport the bug fix to `1.12` (in a new version `1.12.2`)" ]
null
3,450
false
Add `__add__()`, `__iadd__()` and similar to `Dataset` class
**Is your feature request related to a problem? Please describe.** No. **Describe the solution you'd like** I would like to be able to concatenate datasets as follows: ```python >>> dataset["train"] += dataset["validation"] ``` ... instead of using `concatenate_datasets()`: ```python >>> raw_datasets["trai...
https://github.com/huggingface/datasets/issues/3449
[ "I was going through the codebase, and I believe the implementation of __add__() and __iadd__() will be similar to concatenate_datasets() after the elimination of code for arguments other than the list of datasets (info, split, axis). \r\n(Assuming elimination of axis means concatenating over axis 1.)" ]
null
3,449
false
JSONDecodeError with HuggingFace dataset viewer
## Dataset viewer issue for 'pubmed_neg' **Link:** https://huggingface.co/datasets/IGESML/pubmed_neg I am getting the error: Status code: 400 Exception: JSONDecodeError Message: Expecting property name enclosed in double quotes: line 61 column 2 (char 1202) I have checked all files - I am not u...
https://github.com/huggingface/datasets/issues/3448
[ "Hi ! I think the issue comes from the dataset_infos.json file: it has the \"flat\" field twice.\r\n\r\nCan you try deleting this file and regenerating it please ?", "Thanks! That fixed that, but now I am getting:\r\nServer Error\r\nStatus code: 400\r\nException: KeyError\r\nMessage: 'feature'\r\n\r\n...
null
3,448
false