title
stringlengths
1
290
body
stringlengths
0
228k
βŒ€
html_url
stringlengths
46
51
comments
list
pull_request
dict
number
int64
1
5.59k
is_pull_request
bool
2 classes
iter_archive for zip files
* In this PR, I added the option to iterate through zipfiles for `download_manager.py` only. * Next PR will be the same applied to `streaming_download_manager.py`. * Related issue #3272. ## Comments : * There is no `.isreg()` equivalent in zipfile library to check if file is Regular so I used `.is_dir()` instead ...
https://github.com/huggingface/datasets/pull/3347
[ "And also don't always try streaming with Google Drive - it can have issues because of how Google Drive works (with quotas, restrictions, etc.) and it can indeed cause `BlockSizeError`.\r\n\r\nFeel free to host your test data elsewhere, such as in a dataset repository on https://huggingface.co (see [here](https://h...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3347", "html_url": "https://github.com/huggingface/datasets/pull/3347", "diff_url": "https://github.com/huggingface/datasets/pull/3347.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3347.patch", "merged_at": null }
3,347
true
Failed to convert `string` with pyarrow for QED since 1.15.0
## Describe the bug Loading QED was fine until 1.15.0. related: bigscience-workshop/promptsource#659, bigscience-workshop/promptsource#670 Not sure where the root cause is, but here are some candidates: - #3158 - #3120 - #3196 - #2891 ## Steps to reproduce the bug ```python load_dataset("qed") ``` ## ...
https://github.com/huggingface/datasets/issues/3346
[ "Scratch that, probably the old and incompatible usage of dataset builder from promptsource.", "Actually, re-opening this issue cause the error persists\r\n\r\n```python\r\n>>> load_dataset(\"qed\")\r\nDownloading and preparing dataset qed/qed (download: 13.43 MiB, generated: 9.70 MiB, post-processed: Unknown siz...
null
3,346
false
Failed to download species_800 from Google Drive zip file
## Describe the bug One can manually download the zip file on Google Drive, but `load_dataset()` cannot. related: #3248 ## Steps to reproduce the bug ```shell > python Python 3.7.12 (default, Sep 5 2021, 08:34:29) [Clang 11.0.3 (clang-1103.0.32.62)] on darwin Type "help", "copyright", "credits" or "license" ...
https://github.com/huggingface/datasets/issues/3345
[ "Hi,\r\n\r\nthe dataset is downloaded normally on my machine. Maybe the URL was down at the time of your download. Could you try again?", "> Hi,\r\n> \r\n> the dataset is downloaded normally on my machine. Maybe the URL was down at the time of your download. Could you try again?\r\n\r\nI have tried that many time...
null
3,345
false
Add ArrayXD docs
Documents support for dynamic first dimension in `ArrayXD` from #2891, and explain the `ArrayXD` feature in general. Let me know if I'm missing anything @lhoestq :)
https://github.com/huggingface/datasets/pull/3344
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3344", "html_url": "https://github.com/huggingface/datasets/pull/3344", "diff_url": "https://github.com/huggingface/datasets/pull/3344.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3344.patch", "merged_at": "2021-12-01T19:35...
3,344
true
Better error message when download fails
From our discussions in https://github.com/huggingface/datasets/issues/3269 and https://github.com/huggingface/datasets/issues/3282 it would be nice to have better messages if a download fails. In particular the error now shows: - the error from the HEAD request if there's one - otherwise the response code of the ...
https://github.com/huggingface/datasets/pull/3343
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3343", "html_url": "https://github.com/huggingface/datasets/pull/3343", "diff_url": "https://github.com/huggingface/datasets/pull/3343.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3343.patch", "merged_at": "2021-12-01T11:27...
3,343
true
Fix ASSET dataset data URLs
Change the branch name "master" to "main" in the data URLs, since facebookresearch has changed that.
https://github.com/huggingface/datasets/pull/3342
[ "> Hi @tianjianjiang, thanks for the fix.\r\n> The links should also be updated in the `dataset_infos.json` file.\r\n> The failing tests are due to the missing tag in the header of the `README.md` file:\r\n\r\nHi @albertvillanova, thank you for the info! My apologies for the messy PR.\r\n" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3342", "html_url": "https://github.com/huggingface/datasets/pull/3342", "diff_url": "https://github.com/huggingface/datasets/pull/3342.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3342.patch", "merged_at": "2021-12-14T14:50...
3,342
true
Mirror the canonical datasets to the Hugging Face Hub
- [ ] create a repo on https://hf.co/datasets for every canonical dataset - [ ] on every commit related to a dataset, update the hf.co repo See https://github.com/huggingface/moon-landing/pull/1562 @SBrandeis: I let you edit this description if needed to precise the intent.
https://github.com/huggingface/datasets/issues/3341
[ "I created a GitHub project to keep track of what needs to be done:\r\nhttps://github.com/huggingface/datasets/projects/3\r\n\r\nI also store my code in a (private for now) repository at https://github.com/huggingface/mirror_canonical_datasets_on_hub", "I understand that the datasets are mirrored on the Hub now, ...
null
3,341
false
Fix JSON ClassLabel casting for integers
Loading a JSON dataset with ClassLabel feature types currently fails if the JSON data already has integers. Indeed currently it tries to convert the strings to integers without even checking if the data are not integers already. For example this currently fails: ```python from datasets import load_dataset, Feature...
https://github.com/huggingface/datasets/pull/3340
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3340", "html_url": "https://github.com/huggingface/datasets/pull/3340", "diff_url": "https://github.com/huggingface/datasets/pull/3340.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3340.patch", "merged_at": "2021-12-01T11:27...
3,340
true
to_tf_dataset fails on TPU
Using `to_tf_dataset` to create a dataset and then putting it in `model.fit` results in an internal error on TPUs. I've only tried on Colab and Kaggle TPUs, not GCP TPUs. ## Steps to reproduce the bug I made a colab to show the error. https://colab.research.google.com/drive/12x_PFKzGouFxqD4OuWfnycW_1TaT276z?usp=s...
https://github.com/huggingface/datasets/issues/3339
[ "This might be related to https://github.com/tensorflow/tensorflow/issues/38762 , what do you think @Rocketknight1 ?\r\n> Dataset.from_generator is expected to not work with TPUs as it uses py_function underneath which is incompatible with Cloud TPU 2VM setup. If you would like to read from large datasets, maybe tr...
null
3,339
false
[WIP] Add doctests for tutorials
Opening a PR as discussed with @LysandreJik for some help with doctest issues. The goal is to add doctests for each of the tutorials in the documentation to make sure the code samples work as shown. ### Issues A doctest has been added in the docstring of the `load_dataset_builder` function in `load.py` to handle ...
https://github.com/huggingface/datasets/pull/3338
[ "I manage to remove the mentions of ellipsis in the code by launching the command as follows:\r\n\r\n```\r\npython -m doctest -v docs/source/load_hub.rst -o=ELLIPSIS\r\n```\r\n\r\nThe way you put your ellipsis will only work on mac, I've adapted it for linux as well with the following:\r\n\r\n```diff\r\n >>> fro...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3338", "html_url": "https://github.com/huggingface/datasets/pull/3338", "diff_url": "https://github.com/huggingface/datasets/pull/3338.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3338.patch", "merged_at": null }
3,338
true
Typing of Dataset.__getitem__ could be improved.
## Describe the bug The newly added typing for Dataset.__getitem__ is Union[Dict, List]. This makes tools like mypy a bit awkward to use as we need to check the type manually. We could use type overloading to make this easier. [Documentation](https://docs.python.org/3/library/typing.html#typing.overload) ## Steps...
https://github.com/huggingface/datasets/issues/3337
[ "Hi ! Thanks for the suggestion, I didn't know about this decorator.\r\n\r\nIf you are interesting in contributing, feel free to open a pull request to add the overload methods for each typing combination :) To assign you to this issue, you can comment `#self-assign` in this thread.\r\n\r\n`Dataset.__getitem__` is ...
null
3,337
false
Add support for multiple dynamic dimensions and to_pandas conversion for dynamic arrays
Add support for multiple dynamic dimensions (e.g. `(None, None, 3)` for arbitrary sized images) and `to_pandas()` conversion for dynamic arrays. TODOs: * [ ] Cleaner code * [ ] Formatting issues (if NumPy doesn't allow broadcasting even though dtype is np.object) * [ ] Fix some issues with zero-dim tensors * [ ...
https://github.com/huggingface/datasets/pull/3336
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3336", "html_url": "https://github.com/huggingface/datasets/pull/3336", "diff_url": "https://github.com/huggingface/datasets/pull/3336.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3336.patch", "merged_at": null }
3,336
true
add Speech commands dataset
closes #3283
https://github.com/huggingface/datasets/pull/3335
[ "@anton-l ping", "@lhoestq \r\nHi Quentin! Thank you for your feedback and suggestions! πŸ€—\r\n\r\nYes, that was actually what I wanted to do next - I mean the steaming stuff :)\r\nAlso, I need to make some changes to the readme (to account for the updated features set).\r\n\r\nHopefully, I will be done by tomorro...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3335", "html_url": "https://github.com/huggingface/datasets/pull/3335", "diff_url": "https://github.com/huggingface/datasets/pull/3335.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3335.patch", "merged_at": "2021-12-10T10:30...
3,335
true
Integrate Polars library
Check potential integration of the Polars library: https://github.com/pola-rs/polars - Benchmark: https://h2oai.github.io/db-benchmark/ CC: @thomwolf @lewtun
https://github.com/huggingface/datasets/issues/3334
[ "If possible, a neat API could be something like `Dataset.to_polars()`, as well as `Dataset.set_format(\"polars\")`", "Note they use a \"custom\" implementation of Arrow: [Arrow2](https://github.com/jorgecarleitao/arrow2).", "Polars has grown rapidly in popularity over the last year - could you consider integra...
null
3,334
false
load JSON files, get the errors
Hi, does this bug be fixed? when I load JSON files, I get the same errors by the command `!python3 run.py --do_train --task qa --dataset squad-retrain-data/train-v2.0.json --output_dir ./re_trained_model/` change the dateset to load json by refering to https://huggingface.co/docs/datasets/loading.html `dataset = ...
https://github.com/huggingface/datasets/issues/3333
[ "Hi ! The message you're getting is not an error. It simply says that your JSON dataset is being prepared to a location in `/root/.cache/huggingface/datasets`", "> \r\n\r\nbut I want to load local JSON file by command\r\n`python3 run.py --do_train --task qa --dataset squad-retrain-data/train-v2.0.json --output_di...
null
3,333
false
Fix error message and add extension fallback
Fix the error message raised if `infered_module_name` is `None` in `CommunityDatasetModuleFactoryWithoutScript.get_module` and make `infer_module_for_data_files` more robust. In the linked issue, `infer_module_for_data_files` returns `None` because `json` is the second most common extension due to the suffix orderi...
https://github.com/huggingface/datasets/pull/3332
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3332", "html_url": "https://github.com/huggingface/datasets/pull/3332", "diff_url": "https://github.com/huggingface/datasets/pull/3332.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3332.patch", "merged_at": "2021-11-29T13:34...
3,332
true
AttributeError: 'CommunityDatasetModuleFactoryWithoutScript' object has no attribute 'path'
## Describe the bug I add a new question answering dataset to huggingface datasets manually. Here is the link: [luozhouyang/question-answering-datasets](https://huggingface.co/datasets/luozhouyang/question-answering-datasets) But when I load the dataset, an error raised: ```bash AttributeError: 'CommunityDatas...
https://github.com/huggingface/datasets/issues/3331
[ "Hi,\r\n\r\nthe fix was merged and will be available in the next release of `datasets`.\r\nIn the meantime, you can use it by installing `datasets` directly from master as follows:\r\n```\r\npip install git+https://github.com/huggingface/datasets.git\r\n```" ]
null
3,331
false
Change TriviaQA license (#3313)
Fixes (#3313)
https://github.com/huggingface/datasets/pull/3330
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3330", "html_url": "https://github.com/huggingface/datasets/pull/3330", "diff_url": "https://github.com/huggingface/datasets/pull/3330.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3330.patch", "merged_at": "2021-11-29T11:24...
3,330
true
Map function: Type error on iter #999
## Describe the bug Using the map function, it throws a type error on iter #999 Here is the code I am calling: ``` dataset = datasets.load_dataset('squad') dataset['validation'].map(text_numbers_to_int, input_columns=['context'], fn_kwargs={'column': 'context'}) ``` text_numbers_to_int returns the input text ...
https://github.com/huggingface/datasets/issues/3329
[ "Hi, thanks for reporting.\r\n\r\nIt would be really helpful if you could provide the actual code of the `text_numbers_to_int` function so we can reproduce the error.", "```\r\ndef text_numbers_to_int(text, column=\"\"):\r\n \"\"\"\r\n Convert text numbers to int.\r\n\r\n :param text: text numbers\r\n ...
null
3,329
false
Quick fix error formatting
While working on a dataset, I got the error ``` TypeError: Provided `function` which is applied to all elements of table returns a `dict` of types {[type(x) for x in processed_inputs.values()]}. When using `batched=True`, make sure provided `function` returns a `dict` of types like `{allowed_batch_return_types}`. ``...
https://github.com/huggingface/datasets/pull/3328
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3328", "html_url": "https://github.com/huggingface/datasets/pull/3328", "diff_url": "https://github.com/huggingface/datasets/pull/3328.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3328.patch", "merged_at": "2021-11-29T13:32...
3,328
true
"Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)"
## Describe the bug Passing a correctly shaped Numpy-Array to get_nearest_examples leads to the Exception "Shape of query is incorrect, it has to be either a 1D array or 2D (1, N)" Probably the reason for this is a wrongly converted assertion. 1.15.1: `assert len(query.shape) == 1 or (len(query.shape) == 2...
https://github.com/huggingface/datasets/issues/3327
[ "#3323 " ]
null
3,327
false
Fix import `datasets` on python 3.10
In python 3.10 it's no longer possible to use `functools.wraps` on a method decorated with `classmethod`. To fix this I inverted the order of the `inject_arrow_table_documentation` and `classmethod` decorators Fix #3324
https://github.com/huggingface/datasets/pull/3326
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3326", "html_url": "https://github.com/huggingface/datasets/pull/3326", "diff_url": "https://github.com/huggingface/datasets/pull/3326.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3326.patch", "merged_at": "2021-11-26T16:31...
3,326
true
Update conda dependencies
Some dependencies minimum versions were outdated. For example `pyarrow` and `huggingface_hub`
https://github.com/huggingface/datasets/pull/3325
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3325", "html_url": "https://github.com/huggingface/datasets/pull/3325", "diff_url": "https://github.com/huggingface/datasets/pull/3325.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3325.patch", "merged_at": "2021-11-26T16:20...
3,325
true
Can't import `datasets` in python 3.10
When importing `datasets` I'm getting this error in python 3.10: ```python Traceback (most recent call last): File "<string>", line 1, in <module> File "/Users/quentinlhoest/Desktop/hf/nlp/src/datasets/__init__.py", line 34, in <module> from .arrow_dataset import Dataset, concatenate_datasets File "/Use...
https://github.com/huggingface/datasets/issues/3324
[]
null
3,324
false
Fix wrongly converted assert
Seems like this assertion was replaced by an exception but the condition got wrongly converted.
https://github.com/huggingface/datasets/pull/3323
[ "Closes #3327 " ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3323", "html_url": "https://github.com/huggingface/datasets/pull/3323", "diff_url": "https://github.com/huggingface/datasets/pull/3323.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3323.patch", "merged_at": "2021-11-26T16:44...
3,323
true
Add missing tags to XTREME
Add missing tags to the XTREME benchmark for better discoverability.
https://github.com/huggingface/datasets/pull/3322
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3322", "html_url": "https://github.com/huggingface/datasets/pull/3322", "diff_url": "https://github.com/huggingface/datasets/pull/3322.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3322.patch", "merged_at": "2021-11-29T13:40...
3,322
true
Update URL of tatoeba subset of xtreme
Updates the URL of the tatoeba subset of xtreme. Additionally, replaces `os.path.join` with `xjoin` to correctly join the URL segments on Windows. Fix #3320
https://github.com/huggingface/datasets/pull/3321
[ "<s>To be more precise: `os.path.join` is replaced on-the-fly by `xjoin` anyway with patching, to extend it to remote files</s>", "Oh actually just ignore what I said: they were used to concatenate URLs, which is not recommended. Let me fix that again by appending using `+`" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3321", "html_url": "https://github.com/huggingface/datasets/pull/3321", "diff_url": "https://github.com/huggingface/datasets/pull/3321.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3321.patch", "merged_at": "2021-11-26T10:30...
3,321
true
Can't get tatoeba.rus dataset
## Describe the bug It gives an error. > FileNotFoundError: Couldn't find file at https://github.com/facebookresearch/LASER/raw/master/data/tatoeba/v1/tatoeba.rus-eng.rus ## Steps to reproduce the bug ```python data=load_dataset("xtreme","tatoeba.rus", split="validation") ``` ## Solution The library tries...
https://github.com/huggingface/datasets/issues/3320
[]
null
3,320
false
Add push_to_hub docs
Since #3098 it's now possible to upload a dataset on the Hub directly from python using the `push_to_hub` method. I just added a section in the "Upload a dataset to the Hub" tutorial. I kept the section quite simple but let me know if it sounds good to you @LysandreJik @stevhliu :)
https://github.com/huggingface/datasets/pull/3319
[ "Looks good to me! :)\r\n\r\nMaybe we can mention that users can also set the `private` argument if they want to keep their dataset private? It would lead nicely into the next section on Privacy.", "Thanks for your comments, I fixed the capitalization for consistency and added an passage to mention the `private` ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3319", "html_url": "https://github.com/huggingface/datasets/pull/3319", "diff_url": "https://github.com/huggingface/datasets/pull/3319.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3319.patch", "merged_at": "2021-11-25T14:47...
3,319
true
Finish transition to PyArrow 3.0.0
Finish transition to PyArrow 3.0.0 that was started in #3098.
https://github.com/huggingface/datasets/pull/3318
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3318", "html_url": "https://github.com/huggingface/datasets/pull/3318", "diff_url": "https://github.com/huggingface/datasets/pull/3318.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3318.patch", "merged_at": "2021-11-24T15:35...
3,318
true
Add desc parameter to Dataset filter method
**Is your feature request related to a problem? Please describe.** As I was filtering very large datasets I noticed the filter method doesn't have the desc parameter which is available in the map method. Why don't we add a desc parameter to the filter method both for consistency and it's nice to give some feedback to ...
https://github.com/huggingface/datasets/issues/3317
[ "Hi,\r\n\r\n`Dataset.map` allows more generic transforms compared to `Dataset.filter`, which purpose is very specific (to filter examples based on a condition). That's why I don't think we need the `desc` parameter there for consistency. #3196 has added descriptions to the `Dataset` methods that call `.map` intern...
null
3,317
false
Add RedCaps dataset
## Adding a Dataset - **Name:** RedCaps - **Description:** Web-curated image-text data created by the people, for the people - **Paper:** https://arxiv.org/abs/2111.11431 - **Data:** https://redcaps.xyz/ - **Motivation:** Multimodal image-text dataset: 12M+ Image-text pairs Instructions to add a new dataset c...
https://github.com/huggingface/datasets/issues/3316
[]
null
3,316
false
Removing query params for dynamic URL caching
The main use case for this is to make dynamically generated private URLs (like the ones returned by CommonVoice API) compatible with the datasets' caching logic. Usage example: ```python import datasets class CommonVoice(datasets.GeneratorBasedBuilder): def _info(self): return datasets.DatasetInfo...
https://github.com/huggingface/datasets/pull/3315
[ "IMO it makes more sense to have `ignore_url_params` as an attribute of `DownloadConfig` to avoid defining a new argument in `DownloadManger`'s methods.", "@mariosasko that would make sense to me too, but it seems like `DownloadConfig` wasn't intended to be modified from a dataset loading script. @lhoestq wdyt?",...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3315", "html_url": "https://github.com/huggingface/datasets/pull/3315", "diff_url": "https://github.com/huggingface/datasets/pull/3315.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3315.patch", "merged_at": "2021-11-25T14:44...
3,315
true
Adding arg to pass process rank to `map`
This PR adds a `with_rank` argument to `map` that gives the user the possibility to pass the rank of each process to their function. This is mostly designed for multi-GPU map (each process can be sent to a different device thanks to the rank). I've also added tests. I'm putting the PR up so you can check the code, I'll...
https://github.com/huggingface/datasets/pull/3314
[ "Some commits seem to be there twice (made the mistake of rebasing because I wasn't sure whether the doc had changed), is this an issue @lhoestq ?" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3314", "html_url": "https://github.com/huggingface/datasets/pull/3314", "diff_url": "https://github.com/huggingface/datasets/pull/3314.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3314.patch", "merged_at": "2021-11-24T11:54...
3,314
true
TriviaQA License Mismatch
## Describe the bug TriviaQA Webpage at http://nlp.cs.washington.edu/triviaqa/ says they do not own the copyright to the data. However, Huggingface datasets at https://huggingface.co/datasets/trivia_qa mentions that the dataset is released under Apache License Is the License Information on HuggingFace correct?
https://github.com/huggingface/datasets/issues/3313
[ "Hi ! You're completely right, this must be mentioned in the dataset card.\r\nIf you're interesting in contributing, feel free to open a pull request to mention this in the `trivia_qa` dataset card in the \"Licensing Information\" section at https://github.com/huggingface/datasets/blob/master/datasets/trivia_qa/REA...
null
3,313
false
add bl books genre dataset
First of all thanks for the fantastic library/collection of datasets πŸ€— This pull request adds a dataset of metadata from digitised (mostly 19th Century) books from the British Library The [data](https://bl.iro.bl.uk/concern/datasets/1e1ccb46-65b4-4481-b6f8-b8129d5da053) contains various metadata about the books. In...
https://github.com/huggingface/datasets/pull/3312
[ "To fix the CI, feel free to run the `make style` command to format the code.\r\n\r\nThen it also looks like the dummy_data.zip archives are all empty, which makes the tests fail. Can you try regenerating them ? They should have one file inside which is a dummy version of the file at https://bl.iro.bl.uk/downloads/...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3312", "html_url": "https://github.com/huggingface/datasets/pull/3312", "diff_url": "https://github.com/huggingface/datasets/pull/3312.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3312.patch", "merged_at": "2021-12-02T16:07...
3,312
true
Add WebSRC
## Adding a Dataset - **Name:** WebSRC - **Description:** WebSRC is a novel Web-based Structural Reading Comprehension dataset. It consists of 0.44M question-answer pairs, which are collected from 6.5K web pages with corresponding HTML source code, screenshots and metadata. - **Paper:** https://arxiv.org/abs/2101.0...
https://github.com/huggingface/datasets/issues/3311
[]
null
3,311
false
Fatal error condition occurred in aws-c-io
## Describe the bug Fatal error when using the library ## Steps to reproduce the bug ```python from datasets import load_dataset dataset = load_dataset('wikiann', 'en') ``` ## Expected results No fatal errors ## Actual results ``` Fatal error condition occurred in D:\bld\aws-c-io_1633633258269\work\sou...
https://github.com/huggingface/datasets/issues/3310
[ "Hi ! Are you having this issue only with this specific dataset, or it also happens with other ones like `squad` ?", "@lhoestq It happens also on `squad`. It successfully downloads the whole dataset and then crashes on: \r\n\r\n```\r\nFatal error condition occurred in D:\\bld\\aws-c-io_1633633258269\\work\\source...
null
3,310
false
fix: files counted twice in inferred structure
Files were counted twice in a structure like: ``` my_dataset_local_path/ β”œβ”€β”€ README.md └── data/ β”œβ”€β”€ train/ β”‚ β”œβ”€β”€ shard_0.csv β”‚ β”œβ”€β”€ shard_1.csv β”‚ β”œβ”€β”€ shard_2.csv β”‚ └── shard_3.csv └── valid/ β”œβ”€β”€ shard_0.csv └── shard_1.csv ``` The reason is that they were ...
https://github.com/huggingface/datasets/pull/3309
[ "I see it creates some errors in the tests.\r\n\r\nAnother solution if needed is to add something like `data_files = list(set(data_files))` after [this line](https://github.com/huggingface/datasets/blob/8555197a3fe826e98bd0206c2d031c4488c53c5c/src/datasets/data_files.py#L511)", "Hi ! Thanks for the correction :)\...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3309", "html_url": "https://github.com/huggingface/datasets/pull/3309", "diff_url": "https://github.com/huggingface/datasets/pull/3309.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3309.patch", "merged_at": "2021-11-23T17:00...
3,309
true
"dataset_infos.json" missing for chr_en and mc4
## Describe the bug In the repository, every dataset has its metadata in a file called`dataset_infos.json`. But, this file is missing for two datasets: `chr_en` and `mc4`. ## Steps to reproduce the bug Check [chr_en](https://github.com/huggingface/datasets/tree/master/datasets/chr_en) and [mc4](https://github.com/...
https://github.com/huggingface/datasets/issues/3308
[ "Hi ! Thanks for reporting :) \r\nWe can easily add the metadata for `chr_en` IMO, but for mC4 it will take more time, since it requires to count the number of examples in each language", "No problem. I am trying to do some analysis on the metadata of all available datasets. Is reading `metadata_infos.json` for e...
null
3,308
false
Add IndoNLI dataset
This PR adds IndoNLI dataset, from https://aclanthology.org/2021.emnlp-main.821/
https://github.com/huggingface/datasets/pull/3307
[ "@lhoestq thanks for the review! I've modified the labels to follow other NLI datasets.\r\nPlease review my change and let me know if I miss anything." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3307", "html_url": "https://github.com/huggingface/datasets/pull/3307", "diff_url": "https://github.com/huggingface/datasets/pull/3307.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3307.patch", "merged_at": "2021-11-25T14:51...
3,307
true
nested sequence feature won't encode example if the first item of the outside sequence is an empty list
## Describe the bug As the title, nested sequence feature won't encode example if the first item of the outside sequence is an empty list. ## Steps to reproduce the bug ```python from datasets import Features, Sequence, ClassLabel features = Features({ 'x': Sequence(Sequence(ClassLabel(names=['a', 'b']))), ...
https://github.com/huggingface/datasets/issues/3306
[ "knock knock", "Hi, thanks for reporting! I've linked a PR that should fix the issue.", "I've checked the PR and it looks great, thanks a lot!" ]
null
3,306
false
asserts replaced with exception for ``fingerprint.py``, ``search.py``, ``arrow_writer.py`` and ``metric.py``
Addresses #3171 Fixes exception for ``fingerprint.py``, ``search.py``, ``arrow_writer.py`` and ``metric.py`` and modified tests
https://github.com/huggingface/datasets/pull/3305
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3305", "html_url": "https://github.com/huggingface/datasets/pull/3305", "diff_url": "https://github.com/huggingface/datasets/pull/3305.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3305.patch", "merged_at": "2021-11-22T17:08...
3,305
true
Dataset object has no attribute `to_tf_dataset`
I am following HuggingFace Course. I am at Fine-tuning a model. Link: https://huggingface.co/course/chapter3/2?fw=tf I use tokenize_function and `map` as mentioned in the course to process data. `# define a tokenize function` `def Tokenize_function(example):` ` return tokenizer(example['sentence'], truncat...
https://github.com/huggingface/datasets/issues/3304
[ "The issue is due to the older version of transformers and datasets. It has been resolved by upgrading their versions.\r\n\r\n```\r\n# upgrade transformers and datasets to latest versions\r\n!pip install --upgrade transformers\r\n!pip install --upgrade datasets\r\n```\r\n\r\nRegards!" ]
null
3,304
false
DataCollatorWithPadding: TypeError
Hi, I am following the HuggingFace course. I am now at Fine-tuning [https://huggingface.co/course/chapter3/3?fw=tf](https://huggingface.co/course/chapter3/3?fw=tf). When I set up `DataCollatorWithPadding` as following I got an error while trying to reproduce the course code in Kaggle. This error occurs with either a ...
https://github.com/huggingface/datasets/issues/3303
[ "\r\n> \r\n> Input:\r\n> \r\n> ```\r\n> tokenizer = AutoTokenizer.from_pretrained(checkpoint)\r\n> data_collator = DataCollatorWithPadding(tokenizer=tokenizer, return_tensors=\"tf\")\r\n> ```\r\n> \r\n> Output:\r\n> \r\n> ```\r\n> TypeError Traceback (most recent call last)\r\n> /tmp...
null
3,303
false
fix old_val typo in f-string
This PR is to correct a typo in #3277 that @Carlosbogo revieled in a comment. Related closed issue : #3257 Sorry about that πŸ˜….
https://github.com/huggingface/datasets/pull/3302
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3302", "html_url": "https://github.com/huggingface/datasets/pull/3302", "diff_url": "https://github.com/huggingface/datasets/pull/3302.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3302.patch", "merged_at": "2021-11-22T17:04...
3,302
true
Add wikipedia tags
Add the missing tags to the wikipedia dataset card. I also added the missing languages code in our language codes list. This should also fix the code snippet that is presented on the Hub to load the dataset: fix https://github.com/huggingface/datasets/issues/3292
https://github.com/huggingface/datasets/pull/3301
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3301", "html_url": "https://github.com/huggingface/datasets/pull/3301", "diff_url": "https://github.com/huggingface/datasets/pull/3301.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3301.patch", "merged_at": "2021-11-19T16:49...
3,301
true
❓ Dataset loading script from Hugging Face Hub
Hi there, I am trying to add my custom `ag_news` with its own loading script on the Hugging Face datasets hub. In particular, I would like to test the addition of a second configuration to the existing `ag_news` dataset. Once it works in my hub, I plan to make a PR to the original dataset. However, in trying to do s...
https://github.com/huggingface/datasets/issues/3300
[ "Hi ! In the next version of `datasets`, your train and test splits will be correctly separated (changes from #3027) if you create a dataset repository with only your CSV files.\r\n\r\nAlso it seems that you overwrite the `data_files` and `data_dir` arguments in your code, when you instantiate the AGNewsConfig obje...
null
3,300
false
Add option to find unique elements in nested sequences when calling `Dataset.unique`
It would be nice to have an option to flatten nested sequences to find unique elements stored in them when calling `Dataset.unique`. ~~Currently, `Dataset.unique` only supports finding unique sequences and not unique elements in that situation.~~
https://github.com/huggingface/datasets/issues/3299
[ "Hi @mariosasko!\r\n\r\nHas this been patched into any of the releases?", "Hi! Not yet, would you be interested in contributing a PR? I can give you some pointers if needed. " ]
null
3,299
false
Agnews dataset viewer is not working
## Dataset viewer issue for '*name of the dataset*' **Link:** https://huggingface.co/datasets/ag_news Hi there, the `ag_news` dataset viewer is not working. Am I the one who added this dataset? No
https://github.com/huggingface/datasets/issues/3298
[ "Hi ! Thanks for reporting\r\nWe've already fixed the code that generates the preview for this dataset, we'll release the fix soon :)", "Hi @lhoestq, thanks for your feedback!", "Fixed in the viewer.\r\n\r\nhttps://huggingface.co/datasets/ag_news" ]
null
3,298
false
.map() cache is wrongfully reused - only happens when the mapping function is imported
## Describe the bug When `.map` is used with a mapping function that is imported, the cache is reused even if the mapping function has been modified. The reason for this is that `dill` that is used for creating the fingerprint [pickles imported functions by reference](https://stackoverflow.com/a/67851411). I guess...
https://github.com/huggingface/datasets/issues/3297
[ "Hi ! Thanks for reporting. Indeed this is a current limitation of the usage we have of `dill` in `datasets`. I'd suggest you use your workaround for now until we find a way to fix this. Maybe functions that are not coming from a module not installed with pip should be dumped completely, rather than only taking the...
null
3,297
false
Fix temporary dataset_path creation for URIs related to remote fs
This aims to close #3295
https://github.com/huggingface/datasets/pull/3296
[ "Hi ! Thanks for the fix :) \r\n\r\nI think this should be `extract_path_from_uri` 's responsibility to strip the extra `/` from a badly formatted path like `hdfs:///absolute/path` (or raise an error). Do you think you can simply do the changes in `extract_path_from_uri` ? This way this fix will be available for al...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3296", "html_url": "https://github.com/huggingface/datasets/pull/3296", "diff_url": "https://github.com/huggingface/datasets/pull/3296.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3296.patch", "merged_at": "2021-12-06T10:45...
3,296
true
Temporary dataset_path for remote fs URIs not built properly in arrow_dataset.py::load_from_disk
## Describe the bug When trying to build a temporary dataset path from a remote URI in this block of code: https://github.com/huggingface/datasets/blob/42f6b1d18a4a1b6009b6e62d115491be16dfca22/src/datasets/arrow_dataset.py#L1038-L1042 the result is not the expected when passing an absolute path in an URI like `h...
https://github.com/huggingface/datasets/issues/3295
[ "Hi ! Good catch and thanks for opening a PR :)\r\n\r\nI just responded in your PR" ]
null
3,295
false
Add Natural Adversarial Objects dataset
## Adding a Dataset - **Name:** Natural Adversarial Objects (NAO) - **Description:** Natural Adversarial Objects (NAO) is a new dataset to evaluate the robustness of object detection models. NAO contains 7,934 images and 9,943 objects that are unmodified and representative of real-world scenarios, but cause state-of-...
https://github.com/huggingface/datasets/issues/3294
[]
null
3,294
false
Pin version exclusion for Markdown
As Markdown version 3.3.5 has a bug, it is better to exclude it in case the users have it previously installed in their environment. Related to #3289, #3286.
https://github.com/huggingface/datasets/pull/3293
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3293", "html_url": "https://github.com/huggingface/datasets/pull/3293", "diff_url": "https://github.com/huggingface/datasets/pull/3293.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3293.patch", "merged_at": "2021-11-18T10:28...
3,293
true
Not able to load 'wikipedia' dataset
## Describe the bug I am following the instruction for loading the wikipedia dataset using datasets. However getting the below error. ## Steps to reproduce the bug from datasets import load_dataset dataset = load_dataset("wikipedia") ``` ## Expected results A clear and concise description of the expected res...
https://github.com/huggingface/datasets/issues/3292
[ "Hi ! Indeed it looks like the code snippet on the Hugging face Hub doesn't show the second parameter\r\n\r\n![image](https://user-images.githubusercontent.com/42851186/142649237-45ba55c5-1a64-4c30-8692-2c8120572f92.png)\r\n\r\nThanks for reporting, I'm taking a look\r\n" ]
null
3,292
false
Use f-strings in the dataset scripts
Uses f-strings to format the .py files in the dataset folder
https://github.com/huggingface/datasets/pull/3291
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3291", "html_url": "https://github.com/huggingface/datasets/pull/3291", "diff_url": "https://github.com/huggingface/datasets/pull/3291.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3291.patch", "merged_at": "2021-11-22T16:40...
3,291
true
Make several audio datasets streamable
<s>Needs https://github.com/huggingface/datasets/pull/3129 to be merged first</s> Make those audio datasets streamable: - [x] common_voice - [x] openslr - [x] vivos - [x] librispeech_asr <s>(still has some issues to read FLAC)</s> *actually it's ok* - [ ] <s>multilingual_librispeech (yet to be converted)</S> *T...
https://github.com/huggingface/datasets/pull/3290
[ "Reading FLAC (for `librispeech_asr`) works OK for me (`soundfile` version: `0.10.3`):\r\n```python\r\nIn [2]: ds = load_dataset(\"datasets/librispeech_asr/librispeech_asr.py\", \"clean\", streaming=True, split=\"train.100\")\r\n\r\nIn [3]: item = next(iter(ds))\r\n\r\nIn [4]: item.keys()\r\nOut[4]: dict_keys(['fil...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3290", "html_url": "https://github.com/huggingface/datasets/pull/3290", "diff_url": "https://github.com/huggingface/datasets/pull/3290.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3290.patch", "merged_at": "2021-11-19T15:08...
3,290
true
Unpin markdown for build_docs now that it's fixed
`markdown`'s bug has been fixed, so this PR reverts #3286
https://github.com/huggingface/datasets/pull/3289
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3289", "html_url": "https://github.com/huggingface/datasets/pull/3289", "diff_url": "https://github.com/huggingface/datasets/pull/3289.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3289.patch", "merged_at": "2021-11-17T16:23...
3,289
true
Allow datasets with indices table when concatenating along axis=1
Calls `flatten_indices` on the datasets with indices table in `concatenate_datasets` to fix issues when concatenating along `axis=1`. cc @lhoestq: I decided to flatten all the datasets instead of flattening all the datasets except the largest one in the end. The latter approach fails on the following example: ```...
https://github.com/huggingface/datasets/pull/3288
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3288", "html_url": "https://github.com/huggingface/datasets/pull/3288", "diff_url": "https://github.com/huggingface/datasets/pull/3288.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3288.patch", "merged_at": "2021-11-17T15:41...
3,288
true
Add The Pile dataset and PubMed Central subset
Add: - The complete final version of The Pile dataset: "all" config - PubMed Central subset of The Pile: "pubmed_central" config Close #1675, close bigscience-workshop/data_tooling#74. CC: @StellaAthena, @lewtun
https://github.com/huggingface/datasets/pull/3287
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3287", "html_url": "https://github.com/huggingface/datasets/pull/3287", "diff_url": "https://github.com/huggingface/datasets/pull/3287.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3287.patch", "merged_at": "2021-12-01T15:29...
3,287
true
Fix build_docs CI
Because of https://github.com/Python-Markdown/markdown/issues/1196 we have to temporarily pin `markdown` to 3.3.4 for the docs to build without issues
https://github.com/huggingface/datasets/pull/3286
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3286", "html_url": "https://github.com/huggingface/datasets/pull/3286", "diff_url": "https://github.com/huggingface/datasets/pull/3286.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3286.patch", "merged_at": "2021-11-17T11:19...
3,286
true
Add IEMOCAP dataset
## Adding a Dataset - **Name:** IEMOCAP - **Description:** acted, multimodal and multispeaker database - **Paper:** https://sail.usc.edu/iemocap/Busso_2008_iemocap.pdf - **Data:** https://sail.usc.edu/iemocap/index.html - **Motivation:** Useful multimodal dataset cc @anton-l Instructions to add a new datase...
https://github.com/huggingface/datasets/issues/3285
[ "The IEMOCAP dataset is private and available only on request.\r\n```\r\nTo obtain the IEMOCAP data you just need to fill out an electronic release form below.\r\n```\r\n\r\n- [Request form](https://sail.usc.edu/iemocap/release_form.php)\r\n- [License ](https://sail.usc.edu/iemocap/Data_Release_Form_IEMOCAP.pdf)\r\...
null
3,285
false
Add VoxLingua107 dataset
## Adding a Dataset - **Name:** VoxLingua107 - **Description:** VoxLingua107 is a speech dataset for training spoken language identification models. The dataset consists of short speech segments automatically extracted from YouTube videos and labeled according the language of the video title and description, with som...
https://github.com/huggingface/datasets/issues/3284
[ "#self-assign" ]
null
3,284
false
Add Speech Commands dataset
## Adding a Dataset - **Name:** Speech commands - **Description:** A Dataset for Limited-Vocabulary Speech Recognition - **Paper:** https://arxiv.org/abs/1804.03209 - **Data:** https://www.tensorflow.org/datasets/catalog/speech_commands, Available: http://download.tensorflow.org/data/speech_commands_v0.02.tar.gz ...
https://github.com/huggingface/datasets/issues/3283
[ "#self-assign" ]
null
3,283
false
ConnectionError: Couldn't reach https://huggingface.co/datasets/oscar-corpus/OSCAR-2109/resolve/main/OSCAR-2109.py
## Dataset viewer issue for '*oscar-corpus/OSCAR-2109*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/oscar-corpus/OSCAR-2109)* *The dataset library cannot download any language from the oscar-corpus/OSCAR-2109 dataset. By entering the URL in your browser I can access the file.* ...
https://github.com/huggingface/datasets/issues/3282
[ "Hi ! Thanks for reporting :)\r\nI think this is because the dataset is behind an access page. We can fix the dataset viewer\r\n\r\nIf you also have this error when you use the `datasets` library in python, you should probably pass `use_auth_token=True` to the `load_dataset()` function to use your account to access...
null
3,282
false
[Datasets] Improve Covost 2
It's currently quite confusing to understand the manual data download instruction of Covost and not very user-friendly. Currenty the user has to: 1. Go on Common Voice website 2. Find the correct dataset which is **not** mentioned in the error message 3. Download it 4. Untar it 5. Create a language id folder ...
https://github.com/huggingface/datasets/pull/3281
[ "I am trying to use `load_dataset` with the French dataset(common voice corpus 1) which is downloaded from a common voice site and the target language is English (using colab)\r\n\r\nSteps I have followed:\r\n\r\n**1. untar:**\r\n`!tar xvzf fr.tar -C data_dir`\r\n\r\n**2. load data:**\r\n`load_dataset('covost2', 'f...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3281", "html_url": "https://github.com/huggingface/datasets/pull/3281", "diff_url": "https://github.com/huggingface/datasets/pull/3281.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3281.patch", "merged_at": "2021-11-18T10:44...
3,281
true
Fix bookcorpusopen RAM usage
Each document is a full book, so the default arrow writer batch size of 10,000 is too big, and it can fill up RAM quickly before flushing the first batch on disk. I changed its batch size to 256 to use maximum 100MB of memory Fix #3167.
https://github.com/huggingface/datasets/pull/3280
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3280", "html_url": "https://github.com/huggingface/datasets/pull/3280", "diff_url": "https://github.com/huggingface/datasets/pull/3280.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3280.patch", "merged_at": "2021-11-16T13:34...
3,280
true
Minor Typo Fix - Precision to Recall
null
https://github.com/huggingface/datasets/pull/3279
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3279", "html_url": "https://github.com/huggingface/datasets/pull/3279", "diff_url": "https://github.com/huggingface/datasets/pull/3279.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3279.patch", "merged_at": "2021-11-16T11:18...
3,279
true
Proposed update to the documentation for WER
I wanted to submit a minor update to the description of WER for your consideration. Because of the possibility of insertions, the numerator in the WER formula can be larger than N, so the value of WER can be greater than 1.0: ``` >>> from datasets import load_metric >>> metric = load_metric("wer") >>> metric....
https://github.com/huggingface/datasets/pull/3278
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3278", "html_url": "https://github.com/huggingface/datasets/pull/3278", "diff_url": "https://github.com/huggingface/datasets/pull/3278.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3278.patch", "merged_at": "2021-11-16T11:19...
3,278
true
f-string formatting
**Fix #3257** Replaced _.format()_ and _%_ by f-strings in the following modules : - [x] **tests** - [x] **metrics** - [x] **benchmarks** - [x] **utils** - [x] **templates** - [x] **src/Datasets/\*.py** Modules in **_src/Datasets/_**: - [x] **commands** - [x] **features** - [x] **formatting** - [x] **...
https://github.com/huggingface/datasets/pull/3277
[ "Hello @lhoestq, ```make style``` is applied as asked. :)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3277", "html_url": "https://github.com/huggingface/datasets/pull/3277", "diff_url": "https://github.com/huggingface/datasets/pull/3277.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3277.patch", "merged_at": "2021-11-17T16:18...
3,277
true
Update KILT metadata JSON
Fix #3265.
https://github.com/huggingface/datasets/pull/3276
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3276", "html_url": "https://github.com/huggingface/datasets/pull/3276", "diff_url": "https://github.com/huggingface/datasets/pull/3276.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3276.patch", "merged_at": "2021-11-16T11:21...
3,276
true
Force data files extraction if download_mode='force_redownload'
Avoids weird issues when redownloading a dataset due to cached data not being fully updated. With this change, issues #3122 and https://github.com/huggingface/datasets/issues/2956 (not a fix, but a workaround) can be fixed as follows: ```python dset = load_dataset(..., download_mode="force_redownload") ```
https://github.com/huggingface/datasets/pull/3275
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3275", "html_url": "https://github.com/huggingface/datasets/pull/3275", "diff_url": "https://github.com/huggingface/datasets/pull/3275.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3275.patch", "merged_at": "2021-11-15T14:45...
3,275
true
Fix some contact information formats
As reported in https://github.com/huggingface/datasets/issues/3188 some contact information are not displayed correctly. This PR fixes this for CoNLL-2002 and some other datasets with the same issue
https://github.com/huggingface/datasets/pull/3274
[ "The CI fail are caused by some missing sections or tags, which is unrelated to this PR. Merging !" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3274", "html_url": "https://github.com/huggingface/datasets/pull/3274", "diff_url": "https://github.com/huggingface/datasets/pull/3274.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3274.patch", "merged_at": "2021-11-15T14:43...
3,274
true
Respect row ordering when concatenating datasets along axis=1
Currently, there is a bug when concatenating datasets along `axis=1` if more than one dataset has the `_indices` attribute defined. In that scenario, all indices mappings except the first one get ignored. A minimal reproducible example: ```python >>> from datasets import Dataset, concatenate_datasets >>> a = Data...
https://github.com/huggingface/datasets/issues/3273
[]
null
3,273
false
Make iter_archive work with ZIP files
Currently users can use `dl_manager.iter_archive` in their dataset script to iterate over all the files of a TAR archive. It would be nice if it could work with ZIP files too !
https://github.com/huggingface/datasets/issues/3272
[ "Hello, is this issue open for any contributor ? can I work on it ?\r\n\r\n", "Hi ! Sure this is open for any contributor. If you're interested feel free to self-assign this issue to you by commenting `#self-assign`. Then if you have any question or if I can help, feel free to ping me.\r\n\r\nTo begin with, feel ...
null
3,272
false
Decode audio from remote
Currently the Audio feature type can only decode local audio files, not remote files. To fix this I replaced `open` with our `xopen` functoin that is compatible with remote files in audio.py cc @albertvillanova @mariosasko
https://github.com/huggingface/datasets/pull/3271
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3271", "html_url": "https://github.com/huggingface/datasets/pull/3271", "diff_url": "https://github.com/huggingface/datasets/pull/3271.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3271.patch", "merged_at": "2021-11-16T11:35...
3,271
true
Add os.listdir for streaming
Extend `os.listdir` to support streaming data from remote files. This is often used to navigate in remote ZIP files for example
https://github.com/huggingface/datasets/pull/3270
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3270", "html_url": "https://github.com/huggingface/datasets/pull/3270", "diff_url": "https://github.com/huggingface/datasets/pull/3270.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3270.patch", "merged_at": "2021-11-15T10:27...
3,270
true
coqa NonMatchingChecksumError
``` >>> from datasets import load_dataset >>> dataset = load_dataset("coqa") Downloading: 3.82kB [00:00, 1.26MB/s] ...
https://github.com/huggingface/datasets/issues/3269
[ "Hi @ZhaofengWu, thanks for reporting.\r\n\r\nUnfortunately, I'm not able to reproduce your bug:\r\n```python\r\nIn [1]: from datasets import load_dataset\r\n\r\nIn [2]: ds = load_dataset(\"coqa\")\r\nDownloading: 3.82kB [00:00, 1.91MB/s]\r\nDownloading: 1.79kB [00:00, 1.79MB/s]\r\nUsing custom data configuration d...
null
3,269
false
Dataset viewer issue for 'liweili/c4_200m'
## Dataset viewer issue for '*liweili/c4_200m*' **Link:** *[link to the dataset viewer page](https://huggingface.co/datasets/liweili/c4_200m)* *Server Error* ``` Status code: 404 Exception: Status404Error Message: Not found. Maybe the cache is missing, or maybe the ressource does not exist. ``` ...
https://github.com/huggingface/datasets/issues/3268
[ "Hi ! I think the issue comes from this [line](https://huggingface.co/datasets/liweili/c4_200m/blob/main/c4_200m.py#L87):\r\n```python\r\npath = filepath + \"/*.tsv*\"\r\n```\r\n\r\nYou can fix this by doing this instead:\r\n```python\r\npath = os.path.join(filepath, \"/*.tsv*\")\r\n```\r\n\r\nHere is why:\r\n\r\nL...
null
3,268
false
Replacing .format() and % by f-strings
**Fix #3257** Replaced _.format()_ and _%_ by f-strings in the following modules : - [x] **tests** - [x] **metrics** - [x] **benchmarks** - [x] **utils** - [x] **templates** Will follow in the next PR the modules left : - [ ] **src** Module **datasets** will not be edited as asked by @mariosasko PS...
https://github.com/huggingface/datasets/pull/3267
[ "Hi ! It looks like most of your changes are just `black` changes. All those changes are not necessary. In particular if you want to use `black`, please use the `make style` command instead. It runs `black` with additional parameters and you shouldn't end up with that many changes\r\n\r\nFeel free to open a new PR ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3267", "html_url": "https://github.com/huggingface/datasets/pull/3267", "diff_url": "https://github.com/huggingface/datasets/pull/3267.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3267.patch", "merged_at": null }
3,267
true
Fix URLs for WikiAuto Manual, jeopardy and definite_pronoun_resolution
[#3264](https://github.com/huggingface/datasets/issues/3264)
https://github.com/huggingface/datasets/pull/3266
[ "There seems to be problems with datasets metadata, of which I dont have access to. I think one of the datasets is from reddit. Can anyone help?", "Hello @LashaO , I think the errors were caused by `_DATA_FILES` in `definite_pronoun_resolution.py`. Here are details of the test error.\r\n```\r\nself = BuilderConfi...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3266", "html_url": "https://github.com/huggingface/datasets/pull/3266", "diff_url": "https://github.com/huggingface/datasets/pull/3266.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3266.patch", "merged_at": "2021-12-06T11:16...
3,266
true
Checksum error for kilt_task_wow
## Describe the bug Checksum failed when downloads kilt_tasks_wow. See error output for details. ## Steps to reproduce the bug ```python import datasets datasets.load_datasets('kilt_tasks','wow') ``` ## Expected results Download successful ## Actual results ``` Downloading and preparing dataset kilt_ta...
https://github.com/huggingface/datasets/issues/3265
[ "Using `dataset = load_dataset(\"kilt_tasks\", \"wow\", ignore_verifications=True)` may fix it, but I do not think it is a elegant solution.", "Hi @slyviacassell, thanks for reporting.\r\n\r\nYes, there is an issue with the checksum verification. I'm fixing it.\r\n\r\nAnd as you pointed out, in the meantime, you ...
null
3,265
false
Downloading URL change for WikiAuto Manual, jeopardy and definite_pronoun_resolution
## Describe the bug - WikiAuto Manual The original manual datasets with the following downloading URL in this [repository](https://github.com/chaojiang06/wiki-auto) was [deleted](https://github.com/chaojiang06/wiki-auto/commit/0af9b066f2b4e02726fb8a9be49283c0ad25367f) by the author. ``` https://github.com/chaoj...
https://github.com/huggingface/datasets/issues/3264
[ "#take\r\nI am willing to fix this. Links can be replaced for WikiAuto Manual and jeopardy with new ones provided by authors.\r\n\r\nAs for the definite_pronoun_resolution URL, a certificate error seems to be preventing a download. I have the files on my local machine. I can include them in the dataset folder as th...
null
3,264
false
FET DATA
## Adding a Dataset - **Name:** *name of the dataset* - **Description:** *short description of the dataset (or link to social media or blog post)* - **Paper:** *link to the dataset paper if available* - **Data:** *link to the Github repository or current dataset location* - **Motivation:** *what are some good reasons t...
https://github.com/huggingface/datasets/issues/3263
[]
null
3,263
false
asserts replaced with exception for image classification task, csv, json
Fixes for csv, json in io module and image_classification task with tests referenced in https://github.com/huggingface/datasets/issues/3171
https://github.com/huggingface/datasets/pull/3262
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3262", "html_url": "https://github.com/huggingface/datasets/pull/3262", "diff_url": "https://github.com/huggingface/datasets/pull/3262.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3262.patch", "merged_at": "2021-11-15T11:08...
3,262
true
Scifi_TV_Shows: Having trouble getting viewer to find appropriate files
## Dataset viewer issue for '*Science Fiction TV Show Plots Corpus (Scifi_TV_Shows)*' **Link:** [link](https://huggingface.co/datasets/lara-martin/Scifi_TV_Shows) I tried adding both a script (https://huggingface.co/datasets/lara-martin/Scifi_TV_Shows/blob/main/Scifi_TV_Shows.py) and some dummy examples (https:/...
https://github.com/huggingface/datasets/issues/3261
[ "Hi ! I think this is because `iter_archive` doesn't support ZIP files yet. See https://github.com/huggingface/datasets/issues/3272\r\n\r\nYou can navigate into the archive this way instead:\r\n```python\r\n# in split_generators\r\ndata_dir = dl_manager.download_and_extract(url)\r\ntrain_filepath = os.path.join(dat...
null
3,261
false
Fix ConnectionError in Scielo dataset
This PR: * allows 403 status code in HEAD requests to S3 buckets to fix the connection error in the Scielo dataset (instead of `url`, uses `response.url` to check the URL of the final endpoint) * makes the Scielo dataset streamable Fixes #3255.
https://github.com/huggingface/datasets/pull/3260
[ "The CI error is unrelated to the change." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3260", "html_url": "https://github.com/huggingface/datasets/pull/3260", "diff_url": "https://github.com/huggingface/datasets/pull/3260.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3260.patch", "merged_at": "2021-11-16T17:55...
3,260
true
Updating details of IRC disentanglement data
I was pleasantly surprised to find that someone had already added my dataset to the huggingface library, but some details were missing or incorrect. This PR fixes the documentation.
https://github.com/huggingface/datasets/pull/3259
[ "Thank you for the cleanup!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3259", "html_url": "https://github.com/huggingface/datasets/pull/3259", "diff_url": "https://github.com/huggingface/datasets/pull/3259.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3259.patch", "merged_at": "2021-11-18T17:19...
3,259
true
Reload dataset that was already downloaded with `load_from_disk` from cloud storage
`load_from_disk` downloads the dataset to a temporary directory without checking if the dataset has already been downloaded once. It would be nice to have some sort of caching for datasets downloaded this way. This could leverage the fingerprint of the dataset that was saved in the `state.json` file.
https://github.com/huggingface/datasets/issues/3258
[]
null
3,258
false
Use f-strings for string formatting
f-strings offer better readability/performance than `str.format` and `%`, so we should use them in all places in our codebase unless there is good reason to keep the older syntax. > **NOTE FOR CONTRIBUTORS**: To avoid large PRs and possible merge conflicts, do 1-3 modules per PR. Also, feel free to ignore the files ...
https://github.com/huggingface/datasets/issues/3257
[ "Hi, I would be glad to help with this. Is there anyone else working on it?", "Hi, I would be glad to work on this too.", "#self-assign", "Hi @Carlosbogo,\r\n\r\nwould you be interested in replacing the `.format` and `%` syntax with f-strings in the modules in the `datasets` directory since @Mehdi2402 has ope...
null
3,257
false
asserts replaced by exception for text classification task with test.
I have replaced only a single assert in text_classification.py along with a unit test to verify an exception is raised based on https://github.com/huggingface/datasets/issues/3171 . I would like to first understand the code contribution workflow. So keeping the change to a single file rather than making too many ch...
https://github.com/huggingface/datasets/pull/3256
[ "Haha it looks like you got the chance of being reviewed twice at the same time and got the same suggestion twice x)\r\nAnyway it's all good now so we can merge !", "Thanks for the feedback. " ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3256", "html_url": "https://github.com/huggingface/datasets/pull/3256", "diff_url": "https://github.com/huggingface/datasets/pull/3256.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3256.patch", "merged_at": "2021-11-12T14:59...
3,256
true
SciELO dataset ConnectionError
## Describe the bug I get `ConnectionError` when I am trying to load the SciELO dataset. When I try the URL with `requests` I get: ``` >>> requests.head("https://ndownloader.figstatic.com/files/14019287") <Response [302]> ``` And as far as I understand redirections in `datasets` are not supported for downlo...
https://github.com/huggingface/datasets/issues/3255
[]
null
3,255
false
Update xcopa dataset (fix checksum issues + add translated data)
This PR updates the checksums (as reported [here](https://discuss.huggingface.co/t/how-to-load-dataset-locally/11601/2)) of the `xcopa` dataset. Additionally, it adds new configs that hold the translated data of the original set of configs. This data was not available at the time of adding this dataset to the lib.
https://github.com/huggingface/datasets/pull/3254
[ "The CI failures are unrelated to the changes (missing fields in the readme and the CER metric error fixed in #3252)." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3254", "html_url": "https://github.com/huggingface/datasets/pull/3254", "diff_url": "https://github.com/huggingface/datasets/pull/3254.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3254.patch", "merged_at": "2021-11-12T10:30...
3,254
true
`GeneratorBasedBuilder` does not support `None` values
## Describe the bug `GeneratorBasedBuilder` does not support `None` values. ## Steps to reproduce the bug See [this repository](https://github.com/pavel-lexyr/huggingface-datasets-bug-reproduction) for minimal reproduction. ## Expected results Dataset is initialized with a `None` value in the `value` column. ...
https://github.com/huggingface/datasets/issues/3253
[ "Hi,\r\n\r\nthanks for reporting and providing a minimal reproducible example. \r\n\r\nThis line of the PR I've linked in our discussion on the Forum will add support for `None` values:\r\nhttps://github.com/huggingface/datasets/blob/a53de01842aac65c66a49b2439e18fa93ff73ceb/src/datasets/features/features.py#L835\r\...
null
3,253
false
Fix failing CER metric test in CI after update
Fixes the [failing CER metric test](https://app.circleci.com/pipelines/github/huggingface/datasets/8644/workflows/79816553-fa2f-4756-b022-d5937f00bf7b/jobs/53298) in CI by adding support for `jiwer==2.3.0`, which was released yesterday. Also, I verified that all the tests in `metrics/cer/test_cer.py` pass after the cha...
https://github.com/huggingface/datasets/pull/3252
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3252", "html_url": "https://github.com/huggingface/datasets/pull/3252", "diff_url": "https://github.com/huggingface/datasets/pull/3252.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3252.patch", "merged_at": "2021-11-12T14:06...
3,252
true
Add ETHICS dataset
This PR adds the ETHICS dataset, including all 5 sub-datasets. From https://arxiv.org/abs/2008.02275
https://github.com/huggingface/datasets/pull/3250
[ "Thanks for your contribution, @ssss1029. Are you still interested in adding this dataset?\r\n\r\nWe are removing the dataset scripts from this GitHub repo and moving them to the Hugging Face Hub: https://huggingface.co/datasets\r\n\r\nWe would suggest you create this dataset there. Please, feel free to tell us if ...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3250", "html_url": "https://github.com/huggingface/datasets/pull/3250", "diff_url": "https://github.com/huggingface/datasets/pull/3250.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3250.patch", "merged_at": null }
3,250
true
Fix streaming for id_newspapers_2018
To be compatible with streaming, this dataset must use `dl_manager.iter_archive` since the data are in a .tgz file
https://github.com/huggingface/datasets/pull/3249
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3249", "html_url": "https://github.com/huggingface/datasets/pull/3249", "diff_url": "https://github.com/huggingface/datasets/pull/3249.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3249.patch", "merged_at": "2021-11-12T14:01...
3,249
true
Stream from Google Drive and other hosts
Streaming from Google Drive is a bit more challenging than the other host we've been supporting: - the download URL must be updated to add the confirm token obtained by HEAD request - it requires to use cookies to keep the connection alive - the URL doesn't tell any information about whether the file is compressed o...
https://github.com/huggingface/datasets/pull/3248
[ "I just tried some datasets and noticed that `spider` is not working for some reason (the compression type is not recognized), resulting in FileNotFoundError. I can take a look tomorrow", "I'm fixing the remaining files based on TAR archives", "THANKS A LOT" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3248", "html_url": "https://github.com/huggingface/datasets/pull/3248", "diff_url": "https://github.com/huggingface/datasets/pull/3248.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3248.patch", "merged_at": "2021-11-12T17:18...
3,248
true
Loading big json dataset raises pyarrow.lib.ArrowNotImplementedError
## Describe the bug When trying to create a dataset from a json file with around 25MB, the following error is raised `pyarrow.lib.ArrowNotImplementedError: Unsupported cast from struct<b: int64, c: int64> to struct using function cast_struct` Splitting the big file into smaller ones and then loading it with the `lo...
https://github.com/huggingface/datasets/issues/3247
[ "Hi,\r\n\r\nthis issue is similar to https://github.com/huggingface/datasets/issues/3093, so you can either use the solution provided there or try to load the data in one chunk (you can control the chunk size by specifying the `chunksize` parameter (`int`) in `load_dataset`).\r\n\r\n@lhoestq Is this worth opening a...
null
3,247
false