title
stringlengths
1
290
body
stringlengths
0
228k
βŒ€
html_url
stringlengths
46
51
comments
list
pull_request
dict
number
int64
1
5.59k
is_pull_request
bool
2 classes
[tiny] fix typo in stream docs
null
https://github.com/huggingface/datasets/pull/3246
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3246", "html_url": "https://github.com/huggingface/datasets/pull/3246", "diff_url": "https://github.com/huggingface/datasets/pull/3246.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3246.patch", "merged_at": "2021-11-10T11:10...
3,246
true
Fix load_from_disk temporary directory
`load_from_disk` uses `tempfile.TemporaryDirectory()` instead of our `get_temporary_cache_files_directory()` function. This can cause the temporary directory to be deleted before the dataset object is garbage collected. In practice, it prevents anyone from using methods like `shuffle` on a dataset loaded this way, b...
https://github.com/huggingface/datasets/pull/3245
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3245", "html_url": "https://github.com/huggingface/datasets/pull/3245", "diff_url": "https://github.com/huggingface/datasets/pull/3245.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3245.patch", "merged_at": "2021-11-09T15:30...
3,245
true
Fix filter method for batched=True
null
https://github.com/huggingface/datasets/pull/3244
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3244", "html_url": "https://github.com/huggingface/datasets/pull/3244", "diff_url": "https://github.com/huggingface/datasets/pull/3244.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3244.patch", "merged_at": "2021-11-09T15:52...
3,244
true
Remove redundant isort module placement
`isort` can place modules by itself from [version 5.0.0](https://pycqa.github.io/isort/docs/upgrade_guides/5.0.0.html#module-placement-changes-known_third_party-known_first_party-default_section-etc) onwards, making the `known_first_party` and `known_third_party` fields in `setup.cfg` redundant (this is why our CI work...
https://github.com/huggingface/datasets/pull/3243
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3243", "html_url": "https://github.com/huggingface/datasets/pull/3243", "diff_url": "https://github.com/huggingface/datasets/pull/3243.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3243.patch", "merged_at": "2021-11-12T14:02...
3,243
true
Adding ANERcorp-CAMeLLab dataset
null
https://github.com/huggingface/datasets/issues/3242
[ "Adding ANERcorp dataset\r\n\r\n## Adding a Dataset\r\n- **Name:** *ANERcorp-CAMeLLab*\r\n- **Description:** *Since its creation in 2008, the ANERcorp dataset (Benajiba & Rosso, 2008) has been a standard reference used by Arabic named entity recognition researchers around the world. However, over time, this dataset...
null
3,242
false
Swap descriptions of v1 and raw-v1 configs of WikiText dataset and fix metadata
Fix #3237, fix #795.
https://github.com/huggingface/datasets/pull/3241
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3241", "html_url": "https://github.com/huggingface/datasets/pull/3241", "diff_url": "https://github.com/huggingface/datasets/pull/3241.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3241.patch", "merged_at": "2021-11-09T13:49...
3,241
true
Couldn't reach data file for disaster_response_messages
## Describe the bug Following command gives an ConnectionError. ## Steps to reproduce the bug ```python disaster = load_dataset('disaster_response_messages') ``` ## Error ``` ConnectionError: Couldn't reach https://datasets.appen.com/appen_datasets/disaster_response_data/disaster_response_messages_training....
https://github.com/huggingface/datasets/issues/3240
[ "It looks like the dataset isn't available anymore on appen.com\r\n\r\nThe CSV files appear to still be available at https://www.kaggle.com/landlord/multilingual-disaster-response-messages though. It says that the data are under the CC0 license so I guess we can host the dataset elsewhere instead ?" ]
null
3,240
false
Inconsistent performance of the "arabic_billion_words" dataset
## Describe the bug When downloaded from macine 1 the dataset is downloaded and parsed correctly. When downloaded from machine two (which has a different cache directory), the following script: import datasets from datasets import load_dataset raw_dataset_elkhair_1 = load_dataset('arabic_billion_words', 'Alitti...
https://github.com/huggingface/datasets/issues/3239
[]
null
3,239
false
Reuters21578 Couldn't reach
``## Adding a Dataset - **Name:** *Reuters21578* - **Description:** *ConnectionError: Couldn't reach https://kdd.ics.uci.edu/databases/reuters21578/reuters21578.tar.gz* - **Data:** *https://huggingface.co/datasets/reuters21578* `from datasets import load_dataset` `dataset = load_dataset("reuters21578", 'ModLewis...
https://github.com/huggingface/datasets/issues/3238
[ "Hi ! The URL works fine on my side today, could you try again ?", "thank you @lhoestq \r\nit works" ]
null
3,238
false
wikitext description wrong
## Describe the bug Descriptions of the wikitext datasests are wrong. ## Steps to reproduce the bug Please see: https://github.com/huggingface/datasets/blob/f6dcafce996f39b6a4bbe3a9833287346f4a4b68/datasets/wikitext/wikitext.py#L50 ## Expected results The descriptions for raw-v1 and v1 should be switched.
https://github.com/huggingface/datasets/issues/3237
[ "Hi @hongyuanmei, thanks for reporting.\r\n\r\nI'm fixing it.", "Duplicate of:\r\n- #795" ]
null
3,237
false
Loading of datasets changed in #3110 returns no examples
## Describe the bug Loading of datasets changed in https://github.com/huggingface/datasets/pull/3110 returns no examples: ```python DatasetDict({ train: Dataset({ features: ['id', 'title', 'abstract', 'full_text', 'qas'], num_rows: 0 }) validation: Dataset({ features: ['id',...
https://github.com/huggingface/datasets/issues/3236
[ "Hi @eladsegal, thanks for reporting.\r\n\r\nI am sorry, but I can't reproduce the bug:\r\n```\r\nIn [1]: from datasets import load_dataset\r\n\r\nIn [2]: ds = load_dataset(\"qasper\")\r\nDownloading: 5.11kB [00:00, ?B/s]\r\nDownloading and preparing dataset qasper/qasper (download: 9.88 MiB, generated: 35.11 MiB, ...
null
3,236
false
Addd options to use updated bleurt checkpoints
Adds options to use newer recommended checkpoint (as of 2021/10/8) bleurt-20 and its distilled versions. Updated checkpoints are described in https://github.com/google-research/bleurt/blob/master/checkpoints.md#the-recommended-checkpoint-bleurt-20 This change won't affect the default behavior of metrics/bleurt. ...
https://github.com/huggingface/datasets/pull/3235
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3235", "html_url": "https://github.com/huggingface/datasets/pull/3235", "diff_url": "https://github.com/huggingface/datasets/pull/3235.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3235.patch", "merged_at": "2021-11-12T14:05...
3,235
true
Avoid PyArrow type optimization if it fails
Adds a new variable, `DISABLE_PYARROW_TYPES_OPTIMIZATION`, to `config.py` for easier control of the Arrow type optimization. Fix #2206
https://github.com/huggingface/datasets/pull/3234
[ "That's good to have a way to disable this easily :)\r\nI just find it a bit unfortunate that users would have to experience the error once and then do `DISABLE_PYARROW_TYPES_OPTIMIZATION=1`. Do you know if there's a way to simply fallback on disabling it automatically when it fails ?", "@lhoestq Actually, I agre...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3234", "html_url": "https://github.com/huggingface/datasets/pull/3234", "diff_url": "https://github.com/huggingface/datasets/pull/3234.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3234.patch", "merged_at": "2021-11-10T12:04...
3,234
true
Improve repository structure docs
Continuation of the documentation started in https://github.com/huggingface/datasets/pull/3221, taking into account @stevhliu 's comments
https://github.com/huggingface/datasets/pull/3233
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3233", "html_url": "https://github.com/huggingface/datasets/pull/3233", "diff_url": "https://github.com/huggingface/datasets/pull/3233.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3233.patch", "merged_at": "2021-11-09T10:02...
3,233
true
The Xsum datasets seems not able to download.
## Describe the bug The download Link of the Xsum dataset provided in the repository is [Link](http://bollin.inf.ed.ac.uk/public/direct/XSUM-EMNLP18-Summary-Data-Original.tar.gz). It seems not able to download. ## Steps to reproduce the bug ```python load_dataset('xsum') ``` ## Actual results ``` python r...
https://github.com/huggingface/datasets/issues/3232
[ "Hi ! On my side the URL is working fine, could you try again ?", "> Hi ! On my side the URL is working fine, could you try again ?\r\n\r\nI try it again and cannot download the file (might because of my location). Could you please provide another download link(such as google drive)? :>", "I don't know other ...
null
3,232
false
Group tests in multiprocessing workers by test file
By grouping tests by test file, we make sure that all the tests in `test_load.py` are sent to the same worker. Therefore, the fixture `hf_token` will be called only once (and from the same worker). Related to: #3200. Fix #3219.
https://github.com/huggingface/datasets/pull/3231
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3231", "html_url": "https://github.com/huggingface/datasets/pull/3231", "diff_url": "https://github.com/huggingface/datasets/pull/3231.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3231.patch", "merged_at": "2021-11-08T08:59...
3,231
true
Add full tagset to conll2003 README
Even though it is possible to manually get the tagset list with ```python dset.features[field_name].feature.names ``` I think it is useful to have an overview of the used tagset on the dataset card. This is particularly useful in light of the **dataset viewer**: the tags are encoded, so it is not immediately ob...
https://github.com/huggingface/datasets/pull/3230
[ "I also added the missing `pretty_name` tag in the dataset card to fix the CI" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3230", "html_url": "https://github.com/huggingface/datasets/pull/3230", "diff_url": "https://github.com/huggingface/datasets/pull/3230.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3230.patch", "merged_at": "2021-11-09T10:40...
3,230
true
Fix URL in CITATION file
Currently the BibTeX citation parsed from the CITATION file has wrong URL (it shows the repo URL instead of the proceedings paper URL): ``` @inproceedings{Lhoest_Datasets_A_Community_2021, author = {Lhoest, Quentin and Villanova del Moral, Albert and von Platen, Patrick and Wolf, Thomas and Ε aΕ‘ko, Mario and Jernite,...
https://github.com/huggingface/datasets/pull/3229
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3229", "html_url": "https://github.com/huggingface/datasets/pull/3229", "diff_url": "https://github.com/huggingface/datasets/pull/3229.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3229.patch", "merged_at": "2021-11-07T10:04...
3,229
true
Add CITATION file
Add CITATION file.
https://github.com/huggingface/datasets/pull/3228
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3228", "html_url": "https://github.com/huggingface/datasets/pull/3228", "diff_url": "https://github.com/huggingface/datasets/pull/3228.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3228.patch", "merged_at": "2021-11-07T09:51...
3,228
true
Error in `Json(datasets.ArrowBasedBuilder)` class
## Describe the bug When a json file contains a `text` field that is larger than the block_size, the JSON dataset builder fails. ## Steps to reproduce the bug Create a folder that contains the following: ``` . β”œβ”€β”€ testdata β”‚Β Β  └── mydata.json └── test.py ``` Please download [this file](https://github.com/...
https://github.com/huggingface/datasets/issues/3227
[ "I have additionally identified the source of the error, being that [this condition](https://github.com/huggingface/datasets/blob/fc46bba66ba4f432cc10501c16a677112e13984c/src/datasets/packaged_modules/json/json.py#L124-L126) in the file\r\n`python3.8/site-packages/datasets/packaged_modules/json/json.py` is not bein...
null
3,227
false
Fix paper BibTeX citation with proceedings reference
Fix paper BibTeX citation with proceedings reference.
https://github.com/huggingface/datasets/pull/3226
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3226", "html_url": "https://github.com/huggingface/datasets/pull/3226", "diff_url": "https://github.com/huggingface/datasets/pull/3226.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3226.patch", "merged_at": "2021-11-07T07:05...
3,226
true
Update tatoeba to v2021-07-22
Tatoeba's latest version is v2021-07-22
https://github.com/huggingface/datasets/pull/3225
[ "How about this? @lhoestq @abhishekkrthakur ", "Hi ! I think it would be nice if people could still be able to load the old version.\r\nMaybe this can be a parameter ? For example to load the old version they could do\r\n```python\r\nload_dataset(\"tatoeba\", lang1=\"en\", lang2=\"mr\", date=\"v2020-11-09\")\r\n`...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3225", "html_url": "https://github.com/huggingface/datasets/pull/3225", "diff_url": "https://github.com/huggingface/datasets/pull/3225.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3225.patch", "merged_at": "2021-11-12T11:13...
3,225
true
User-pickling with dynamic sub-classing
This is a continuation of the now closed PR in https://github.com/huggingface/datasets/pull/3206. The discussion there has shaped a new approach to do this. In this PR, behavior of `pklregister` and `Pickler` is extended. Earlier, users were already able to register custom pickle functions. That is useful if they ha...
https://github.com/huggingface/datasets/pull/3224
[ "@lhoestq Feel free to have a look. The implementation is slightly different from what you suggested. I have opted to overwrite `save` instead of meddling with `save_global`. `save_global` is called very late down in dill/pickle so it is hard to control for what is happening there. I might be wrong. Pickling is mor...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3224", "html_url": "https://github.com/huggingface/datasets/pull/3224", "diff_url": "https://github.com/huggingface/datasets/pull/3224.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3224.patch", "merged_at": null }
3,224
true
Update BibTeX entry
Update BibTeX entry.
https://github.com/huggingface/datasets/pull/3223
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3223", "html_url": "https://github.com/huggingface/datasets/pull/3223", "diff_url": "https://github.com/huggingface/datasets/pull/3223.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3223.patch", "merged_at": "2021-11-06T07:06...
3,223
true
Add docs for audio processing
This PR adds documentation for the `Audio` feature. It describes: - The difference between loading `path` and `audio`, as well as use-cases/best practices for each of them. - Resampling audio files with `cast_column`, and then calling `ds[0]["audio"]` to automatically decode and resample to the desired sampling rat...
https://github.com/huggingface/datasets/pull/3222
[ "Nice ! love it this way. I guess you can set this PR to \"ready for review\" ?", "I guess we can merge this one now :)" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3222", "html_url": "https://github.com/huggingface/datasets/pull/3222", "diff_url": "https://github.com/huggingface/datasets/pull/3222.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3222.patch", "merged_at": "2021-11-24T15:35...
3,222
true
Resolve data_files by split name
As discussed in https://github.com/huggingface/datasets/issues/3027 we should automatically infer what file is supposed to go to what split automatically, based on filenames. I added the support for different kinds of patterns, for both dataset repositories and local directories: ``` Input structure: ...
https://github.com/huggingface/datasets/pull/3221
[ "Really cool!\r\nWhen splitting by folder, what do we use for validation set (\"valid\", \"validation\" or both)?", "> When splitting by folder, what do we use for validation set (\"valid\", \"validation\" or both)?\r\n\r\nBoth are fine :) As soon as it has \"valid\" in it", "Merging for now, if you have commen...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3221", "html_url": "https://github.com/huggingface/datasets/pull/3221", "diff_url": "https://github.com/huggingface/datasets/pull/3221.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3221.patch", "merged_at": "2021-11-05T17:49...
3,221
true
Add documentation about dataset viewer feature
Add to the docs more details about the dataset viewer feature in the Hub. CC: @julien-c
https://github.com/huggingface/datasets/issues/3220
[]
null
3,220
false
Eventual Invalid Token Error at setup of private datasets
## Describe the bug From time to time, there appear Invalid Token errors with private datasets: - https://app.circleci.com/pipelines/github/huggingface/datasets/8520/workflows/d44629f2-4749-40f8-a657-50931d0b3434/jobs/52534 ``` ____________ ERROR at setup of test_load_streaming_private_dataset _____________ ...
https://github.com/huggingface/datasets/issues/3219
[]
null
3,219
false
Fix code quality in riddle_sense dataset
Fix trailing whitespace. Fix #3217.
https://github.com/huggingface/datasets/pull/3218
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3218", "html_url": "https://github.com/huggingface/datasets/pull/3218", "diff_url": "https://github.com/huggingface/datasets/pull/3218.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3218.patch", "merged_at": "2021-11-04T17:50...
3,218
true
Fix code quality bug in riddle_sense dataset
## Describe the bug ``` datasets/riddle_sense/riddle_sense.py:36:21: W291 trailing whitespace ```
https://github.com/huggingface/datasets/issues/3217
[ "To give more context: https://github.com/psf/black/issues/318. `black` doesn't treat this as a bug, but `flake8` does. \r\n" ]
null
3,217
false
Pin version exclusion for tensorflow incompatible with keras
Once `tensorflow` version 2.6.2 is released: - https://github.com/tensorflow/tensorflow/commit/c1867f3bfdd1042f694df7a9870be51ba80543cb - https://pypi.org/project/tensorflow/2.6.2/ with the patch: - tensorflow/tensorflow#52927 we can remove the temporary fix we introduced in: - #3208 Fix #3209.
https://github.com/huggingface/datasets/pull/3216
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3216", "html_url": "https://github.com/huggingface/datasets/pull/3216", "diff_url": "https://github.com/huggingface/datasets/pull/3216.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3216.patch", "merged_at": "2021-11-05T10:57...
3,216
true
Small updates to to_tf_dataset documentation
I added a little more description about `to_tf_dataset` compared to just setting the format
https://github.com/huggingface/datasets/pull/3215
[ "@stevhliu Accepted both suggestions, thanks for the review!" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3215", "html_url": "https://github.com/huggingface/datasets/pull/3215", "diff_url": "https://github.com/huggingface/datasets/pull/3215.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3215.patch", "merged_at": "2021-11-04T18:55...
3,215
true
Add ACAV100M Dataset
## Adding a Dataset - **Name:** *ACAV100M* - **Description:** *contains 100 million videos with high audio-visual correspondence, ideal for self-supervised video representation learning.* - **Paper:** *https://arxiv.org/abs/2101.10803* - **Data:** *https://github.com/sangho-vision/acav100m* - **Motivation:** *The ...
https://github.com/huggingface/datasets/issues/3214
[]
null
3,214
false
Fix tuple_ie download url
Fix #3204
https://github.com/huggingface/datasets/pull/3213
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3213", "html_url": "https://github.com/huggingface/datasets/pull/3213", "diff_url": "https://github.com/huggingface/datasets/pull/3213.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3213.patch", "merged_at": "2021-11-05T14:16...
3,213
true
Sort files before loading
When loading a dataset that consists of several files (e.g. `my_data/data_001.json`, `my_data/data_002.json` etc.) they are not loaded in order when using `load_dataset("my_data")`. This could lead to counter-intuitive results if, for example, the data files are sorted by date or similar since they would appear in d...
https://github.com/huggingface/datasets/issues/3212
[ "This will be fixed by https://github.com/huggingface/datasets/pull/3221" ]
null
3,212
false
Fix disable_nullable default value to False
Currently the `disable_nullable` parameter is not consistent across all dataset transforms. For example it is `False` in `map` but `True` in `flatten_indices`. This creates unexpected behaviors like this ```python from datasets import Dataset, concatenate_datasets d1 = Dataset.from_dict({"a": [0, 1, 2, 3]}) d2...
https://github.com/huggingface/datasets/pull/3211
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3211", "html_url": "https://github.com/huggingface/datasets/pull/3211", "diff_url": "https://github.com/huggingface/datasets/pull/3211.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3211.patch", "merged_at": "2021-11-04T11:08...
3,211
true
ConnectionError: Couldn't reach https://raw.githubusercontent.com/huggingface/datasets/1.15.1/datasets/wmt16/wmt16.py
when I use python examples/pytorch/translation/run_translation.py --model_name_or_path examples/pytorch/translation/opus-mt-en-ro --do_train --do_eval --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config_name ro-en --output_dir /tmp/tst-translation --per_device_tra...
https://github.com/huggingface/datasets/issues/3210
[ "Hi ! Do you have some kind of proxy in your browser that gives you access to internet ?\r\n\r\nMaybe you're having this error because you don't have access to this URL from python ?", "Hi,do you fixed this error?\r\nI still have this issue when use \"use_auth_token=True\"", "You don't need authentication to ac...
null
3,210
false
Unpin keras once TF fixes its release
Related to: - #3208
https://github.com/huggingface/datasets/issues/3209
[]
null
3,209
false
Pin keras version until TF fixes its release
Fix #3207.
https://github.com/huggingface/datasets/pull/3208
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3208", "html_url": "https://github.com/huggingface/datasets/pull/3208", "diff_url": "https://github.com/huggingface/datasets/pull/3208.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3208.patch", "merged_at": "2021-11-04T09:30...
3,208
true
CI error: Another metric with the same name already exists in Keras 2.7.0
## Describe the bug Release of TensorFlow 2.7.0 contains an incompatibility with Keras. See: - keras-team/keras#15579 This breaks our CI test suite: https://app.circleci.com/pipelines/github/huggingface/datasets/8493/workflows/055c7ae2-43bc-49b4-9f11-8fc71f35a25c/jobs/52363
https://github.com/huggingface/datasets/issues/3207
[]
null
3,207
false
[WIP] Allow user-defined hash functions via a registry
Inspired by the discussion on hashing in https://github.com/huggingface/datasets/issues/3178#issuecomment-959016329, @lhoestq suggested that it would be neat to allow users more control over the hashing process. Specifically, it would be great if users can specify specific hashing functions depending on the **class** o...
https://github.com/huggingface/datasets/pull/3206
[ "Hi @BramVanroy, thanks for your PR.\r\n\r\nThere was a bug in TensorFlow/Keras. We have made a temporary fix in master branch. Please, merge master into your PR branch, so that the CI tests pass.\r\n\r\n```\r\ngit checkout registry\r\ngit fetch upstream master\r\ngit merge upstream/master\r\n```", "@albertvillan...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3206", "html_url": "https://github.com/huggingface/datasets/pull/3206", "diff_url": "https://github.com/huggingface/datasets/pull/3206.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3206.patch", "merged_at": null }
3,206
true
Add Multidoc2dial Dataset
This PR adds the MultiDoc2Dial dataset introduced in this [paper](https://arxiv.org/pdf/2109.12595v1.pdf )
https://github.com/huggingface/datasets/pull/3205
[ "@songfeng cc", "Hi @sivasankalpp, thanks for your PR.\r\n\r\nThere was a bug in TensorFlow/Keras. We have made a temporary fix in our master branch. Please, merge master into your PR branch, so that the CI tests pass.\r\n\r\n```\r\ngit checkout multidoc2dial\r\ngit fetch upstream master\r\ngit merge upstream/mas...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3205", "html_url": "https://github.com/huggingface/datasets/pull/3205", "diff_url": "https://github.com/huggingface/datasets/pull/3205.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3205.patch", "merged_at": "2021-11-24T16:55...
3,205
true
FileNotFoundError for TupleIE dataste
Hi, `dataset = datasets.load_dataset('tuple_ie', 'all')` returns a FileNotFound error. Is the data not available? Many thanks.
https://github.com/huggingface/datasets/issues/3204
[ "@mariosasko @lhoestq Could you give me an update on how to load the dataset after the fix?\r\nThanks.", "Hi @arda-vianai,\r\n\r\nfirst, you can try:\r\n```python\r\nimport datasets\r\ndataset = datasets.load_dataset('tuple_ie', 'all', revision=\"master\")\r\n```\r\nIf this doesn't work, your version of `datasets...
null
3,204
false
Updated: DaNE - updated URL for download
It seems that DaNLP has updated their download URLs and it therefore also needs to be updated in here...
https://github.com/huggingface/datasets/pull/3203
[ "Actually it looks like the old URL is still working, and it's also the one that is mentioned in https://github.com/alexandrainst/danlp/blob/master/docs/docs/datasets.md\r\n\r\nWhat makes you think we should use the new URL ?", "@lhoestq Sorry! I might have jumped to conclusions a bit too fast here... \r\n\r\nI w...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3203", "html_url": "https://github.com/huggingface/datasets/pull/3203", "diff_url": "https://github.com/huggingface/datasets/pull/3203.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3203.patch", "merged_at": "2021-11-04T11:46...
3,203
true
Add mIoU metric
**Is your feature request related to a problem? Please describe.** Recently, some semantic segmentation models were added to HuggingFace Transformers, including [SegFormer](https://huggingface.co/transformers/model_doc/segformer.html) and [BEiT](https://huggingface.co/transformers/model_doc/beit.html). Semantic seg...
https://github.com/huggingface/datasets/issues/3202
[ "Resolved via https://github.com/huggingface/datasets/pull/3745." ]
null
3,202
false
Add GSM8K dataset
## Adding a Dataset - **Name:** GSM8K (short for Grade School Math 8k) - **Description:** GSM8K is a dataset of 8.5K high quality linguistically diverse grade school math word problems created by human problem writers. - **Paper:** https://openai.com/blog/grade-school-math/ - **Data:** https://github.com/openai/gra...
https://github.com/huggingface/datasets/issues/3201
[ "Closed via https://github.com/huggingface/datasets/pull/4103" ]
null
3,201
false
Catch token invalid error in CI
The staging back end sometimes returns invalid token errors when trying to delete a repo. I modified the fixture in the test that uses staging to ignore this error
https://github.com/huggingface/datasets/pull/3200
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3200", "html_url": "https://github.com/huggingface/datasets/pull/3200", "diff_url": "https://github.com/huggingface/datasets/pull/3200.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3200.patch", "merged_at": "2021-11-03T09:41...
3,200
true
Bump huggingface_hub
huggingface_hub just released its first minor version, so we need to update the dependency It was supposed to be part of 1.15.0 but I'm adding it for 1.15.1
https://github.com/huggingface/datasets/pull/3199
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3199", "html_url": "https://github.com/huggingface/datasets/pull/3199", "diff_url": "https://github.com/huggingface/datasets/pull/3199.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3199.patch", "merged_at": "2021-11-02T21:41...
3,199
true
Add Multi-Lingual LibriSpeech
Add https://www.openslr.org/94/
https://github.com/huggingface/datasets/pull/3198
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3198", "html_url": "https://github.com/huggingface/datasets/pull/3198", "diff_url": "https://github.com/huggingface/datasets/pull/3198.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3198.patch", "merged_at": "2021-11-04T17:09...
3,198
true
Fix optimized encoding for arrays
Hi ! #3124 introduced a regression that made the benchmarks CI fail because of a bad array comparison when checking the first encoded element. This PR fixes this by making sure that encoding is applied on all sequence types except lists. cc @eladsegal fyi (no big deal)
https://github.com/huggingface/datasets/pull/3197
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3197", "html_url": "https://github.com/huggingface/datasets/pull/3197", "diff_url": "https://github.com/huggingface/datasets/pull/3197.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3197.patch", "merged_at": "2021-11-02T19:12...
3,197
true
QOL improvements: auto-flatten_indices and desc in map calls
This PR: * automatically calls `flatten_indices` where needed: in `unique` and `save_to_disk` to avoid saving the indices file * adds descriptions to the map calls Fix #3040
https://github.com/huggingface/datasets/pull/3196
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3196", "html_url": "https://github.com/huggingface/datasets/pull/3196", "diff_url": "https://github.com/huggingface/datasets/pull/3196.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3196.patch", "merged_at": "2021-11-02T15:41...
3,196
true
More robust `None` handling
PyArrow has explicit support for `null` values, so it makes sense to support Nones on our side as well. [Colab Notebook with examples](https://colab.research.google.com/drive/1zcK8BnZYnRe3Ao2271u1T19ag9zLEiy3?usp=sharing) Changes: * allow None for the features types with special encoding (`ClassLabel, Translatio...
https://github.com/huggingface/datasets/pull/3195
[ "I also created a PR regarding `disable_nullable` that must be always `False` by default, in order to always allow None values\r\nhttps://github.com/huggingface/datasets/pull/3211", "@lhoestq I addressed your comments, added tests, did some refactoring to make the implementation cleaner and added support for `Non...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3195", "html_url": "https://github.com/huggingface/datasets/pull/3195", "diff_url": "https://github.com/huggingface/datasets/pull/3195.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3195.patch", "merged_at": "2021-12-09T14:26...
3,195
true
Update link to Datasets Tagging app in Spaces
Fix #3193.
https://github.com/huggingface/datasets/pull/3194
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3194", "html_url": "https://github.com/huggingface/datasets/pull/3194", "diff_url": "https://github.com/huggingface/datasets/pull/3194.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3194.patch", "merged_at": "2021-11-08T10:36...
3,194
true
Update link to datasets-tagging app
Once datasets-tagging has been transferred to Spaces: - huggingface/datasets-tagging#22 We should update the link in Datasets.
https://github.com/huggingface/datasets/issues/3193
[]
null
3,193
false
Multiprocessing filter/map (tests) not working on Windows
While running the tests, I found that the multiprocessing examples fail on Windows, or rather they do not complete: they cause a deadlock. I haven't dug deep into it, but they do not seem to work as-is. I currently have no time to tests this in detail but at least the tests seem not to run correctly (deadlocking). #...
https://github.com/huggingface/datasets/issues/3192
[]
null
3,192
false
Dataset viewer issue for '*compguesswhat*'
## Dataset viewer issue for '*compguesswhat*' **Link:** https://huggingface.co/datasets/compguesswhat File not found Am I the one who added this dataset ? No
https://github.com/huggingface/datasets/issues/3191
[ "```python\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('compguesswhat', name='compguesswhat-original',split='train', streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datasets-preview-backend/.ve...
null
3,191
false
combination of shuffle and filter results in a bug
## Describe the bug Hi, I would like to shuffle a dataset, then filter it based on each existing label. however, the combination of `filter`, `shuffle` seems to results in a bug. In the minimal example below, as you see in the filtered results, the filtered labels are not unique, meaning filter has not worked. Any su...
https://github.com/huggingface/datasets/issues/3190
[ "I cannot reproduce this on master and pyarrow==4.0.1.\r\n", "Hi ! There was a regression in `datasets` 1.12 that introduced this bug. It has been fixed in #3019 in 1.13\r\n\r\nCan you try to update `datasets` and try again ?", "Thanks a lot, fixes with 1.13" ]
null
3,190
false
conll2003 incorrect label explanation
In the [conll2003](https://huggingface.co/datasets/conll2003#data-fields) README, the labels are described as follows > - `id`: a `string` feature. > - `tokens`: a `list` of `string` features. > - `pos_tags`: a `list` of classification labels, with possible values including `"` (0), `''` (1), `#` (2), `$` (3), `(`...
https://github.com/huggingface/datasets/issues/3189
[ "Hi @BramVanroy,\r\n\r\nsince these fields are of type `ClassLabel` (you can check this with `dset.features`), you can inspect the possible values with:\r\n```python\r\ndset.features[field_name].feature.names # .feature because it's a sequence of labels\r\n```\r\n\r\nand to find the mapping between names and integ...
null
3,189
false
conll2002 issues
**Link:** https://huggingface.co/datasets/conll2002 The dataset viewer throws a server error when trying to preview the dataset. ``` Message: Extraction protocol 'train' for file at 'https://raw.githubusercontent.com/teropa/nlp/master/resources/corpora/conll2002/esp.train' is not implemented yet ``` I...
https://github.com/huggingface/datasets/issues/3188
[ "Hi ! Thanks for reporting :)\r\n\r\nThis is related to https://github.com/huggingface/datasets/issues/2742, I'm working on it. It should fix the viewer for around 80 datasets.\r\n", "Ah, hadn't seen that sorry.\r\n\r\nThe scrambled \"point of contact\" is a separate issue though, I think.", "@lhoestq The \"poi...
null
3,188
false
Add ChrF(++) (as implemented in sacrebleu)
Similar to my [PR for TER](https://github.com/huggingface/datasets/pull/3153), it feels only right to also include ChrF and friends. These are present in Sacrebleu and are therefore very similar to implement as TER and sacrebleu. I tested the implementation with sacrebleu's tests to verify. You can try this below for y...
https://github.com/huggingface/datasets/pull/3187
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3187", "html_url": "https://github.com/huggingface/datasets/pull/3187", "diff_url": "https://github.com/huggingface/datasets/pull/3187.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3187.patch", "merged_at": "2021-11-02T14:31...
3,187
true
Dataset viewer for nli_tr
## Dataset viewer issue for '*nli_tr*' **Link:** https://huggingface.co/datasets/nli_tr Hello, Thank you for the new dataset preview feature that will help the users to view the datasets online. We just noticed that the dataset viewer widget in the `nli_tr` dataset shows the error below. The error must be d...
https://github.com/huggingface/datasets/issues/3186
[ "It's an issue with the streaming mode:\r\n\r\n```python\r\n>>> import datasets\r\n>>> dataset = datasets.load_dataset('nli_tr', name='snli_tr',split='test', streaming=True)\r\n>>> next(iter(dataset))\r\nTraceback (most recent call last):\r\n File \"<stdin>\", line 1, in <module>\r\n File \"/home/slesage/hf/datas...
null
3,186
false
7z dataset preview not implemented?
## Dataset viewer issue for dataset 'samsum' **Link:** https://huggingface.co/datasets/samsum Server Error Status code: 400 Exception: NotImplementedError Message: Extraction protocol '7z' for file at 'https://arxiv.org/src/1911.12237v2/anc/corpus.7z' is not implemented yet
https://github.com/huggingface/datasets/issues/3185
[ "It's a bug in the dataset viewer: the dataset cannot be downloaded in streaming mode, but since the dataset is relatively small, the dataset viewer should have fallback to normal mode. Working on a fix.", "Fixed. https://huggingface.co/datasets/samsum/viewer/samsum/train\r\n\r\n<img width=\"1563\" alt=\"Capture ...
null
3,185
false
RONEC v2
Hi, as we've recently finished with the new RONEC (Romanian Named Entity Corpus), we'd like to update the dataset here as well. It's actually essential as links to V1 are no longer valid. In reality we'd like to replace completely v1, as v2 is a full re-annotation of v1 with additional data (up to 2x size vs v1). ...
https://github.com/huggingface/datasets/pull/3184
[ "@lhoestq Thanks for the review. I totally understand what you are saying. Normally, I would definitely agree with you, but in this particular case, the quality of v1 is poor, and the dataset itself is small (at the time we created v1 it was the only RO NER dataset, and its size was limited by the available resourc...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3184", "html_url": "https://github.com/huggingface/datasets/pull/3184", "diff_url": "https://github.com/huggingface/datasets/pull/3184.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3184.patch", "merged_at": "2021-11-02T16:02...
3,184
true
Add missing docstring to DownloadConfig
Document the `use_etag` and `num_proc` attributes in `DownloadConig`.
https://github.com/huggingface/datasets/pull/3183
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3183", "html_url": "https://github.com/huggingface/datasets/pull/3183", "diff_url": "https://github.com/huggingface/datasets/pull/3183.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3183.patch", "merged_at": "2021-11-02T10:25...
3,183
true
Don't memoize strings when hashing since two identical strings may have different python ids
When hashing an object that has several times the same string, the hashing could return a different hash if the identical strings share the same python `id()` or not. Here is an example code that shows how the issue can affect the caching: ```python import json import pyarrow as pa from datasets.features import ...
https://github.com/huggingface/datasets/pull/3182
[ "This change slows down the hash computation a little bit but from my tests it doesn't look too impactful. So I think it's fine to merge this." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3182", "html_url": "https://github.com/huggingface/datasets/pull/3182", "diff_url": "https://github.com/huggingface/datasets/pull/3182.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3182.patch", "merged_at": "2021-11-02T09:35...
3,182
true
`None` converted to `"None"` when loading a dataset
## Describe the bug When loading a dataset `None` values of the type `NoneType` are converted to `'None'` of the type `str`. ## Steps to reproduce the bug ```python from datasets import load_dataset qasper = load_dataset("qasper", split="train", download_mode="reuse_cache_if_exists") print(qasper[60]["full_text...
https://github.com/huggingface/datasets/issues/3181
[ "Hi @eladsegal, thanks for reporting.\r\n\r\n@mariosasko I saw you are already working on this, but maybe my comment will be useful to you.\r\n\r\nAll values are casted to their corresponding feature type (including `None` values). For example if the feature type is `Value(\"bool\")`, `None` is casted to `False`.\r...
null
3,181
false
fix label mapping
Fixing label mapping for hlgd. 0 correponds to same event and 1 corresponds to different event <img width="642" alt="Capture d’écran 2021-10-29 aΜ€ 10 39 58 AM" src="https://user-images.githubusercontent.com/16107619/139454810-1f225e3d-ad48-44a8-b8b1-9205c9533839.png"> <img width="638" alt="Capture d’écran 2021-10-...
https://github.com/huggingface/datasets/pull/3180
[ "heck, test failings. moving to draft. will come back to this later today hopefully", "Thanks for fixing this :)\r\nI just updated the dataset_infos.json and added the missing `pretty_name` tag to the dataset card", "thank you @lhoestq! running around as always it felt through as a lower priority..." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3180", "html_url": "https://github.com/huggingface/datasets/pull/3180", "diff_url": "https://github.com/huggingface/datasets/pull/3180.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3180.patch", "merged_at": "2021-11-02T10:37...
3,180
true
Cannot load dataset when the config name is "special"
## Describe the bug After https://github.com/huggingface/datasets/pull/3159, we can get the config name of "Check/region_1", which is "Check___region_1". But now we cannot load the dataset (not sure it's related to the above PR though). It's the case for all the similar datasets, listed in https://github.com/hugg...
https://github.com/huggingface/datasets/issues/3179
[ "The issue is that the datasets are malformed. Not a bug with the datasets library" ]
null
3,179
false
"Property couldn't be hashed properly" even though fully picklable
## Describe the bug I am trying to tokenize a dataset with spaCy. I found that no matter what I do, the spaCy language object (`nlp`) prevents `datasets` from pickling correctly - or so the warning says - even though manually pickling is no issue. It should not be an issue either, since spaCy objects are picklable. ...
https://github.com/huggingface/datasets/issues/3178
[ "After some digging, I found that this is caused by `dill` and using `recurse=True)` when trying to dump the object. The problem also occurs without multiprocessing. I can only find [the following information](https://dill.readthedocs.io/en/latest/dill.html#dill._dill.dumps) about this:\r\n\r\n> If recurse=True, th...
null
3,178
false
More control over TQDM when using map/filter with multiple processes
It would help with the clutter in my terminal if tqdm is only shown for rank 0 when using `num_proces>0` in the map and filter methods of datasets. ```python dataset.map(lambda examples: tokenize(examples["text"]), batched=True, num_proc=6) ``` The above snippet leads to a lot of TQDM bars and depending on your...
https://github.com/huggingface/datasets/issues/3177
[ "Hi,\r\n\r\nIt's hard to provide an API that would cover all use-cases with tqdm in this project.\r\n\r\nHowever, you can make it work by defining a custom decorator (a bit hacky tho) as follows:\r\n```python\r\nimport datasets\r\n\r\ndef progress_only_on_rank_0(func):\r\n def wrapper(*args, **kwargs):\r\n ...
null
3,177
false
OpenSLR dataset: update generate_examples to properly extract data for SLR83
Fixed #3168. The SLR38 indices are CSV files and there wasn't any code in openslr.py to process these files properly. The end result was an empty table. I've added code to properly process these CSV files.
https://github.com/huggingface/datasets/pull/3176
[ "Also fix #3125." ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3176", "html_url": "https://github.com/huggingface/datasets/pull/3176", "diff_url": "https://github.com/huggingface/datasets/pull/3176.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3176.patch", "merged_at": "2021-10-29T10:04...
3,176
true
Add docs for `to_tf_dataset`
This PR adds some documentation for new features released in v1.13.0, with the main addition being `to_tf_dataset`: - Show how to use `to_tf_dataset` in the tutorial, and move `set_format(type='tensorflow'...)` to the Process section (let me know if I'm missing anything @Rocketknight1 πŸ˜…). - Add an example for load...
https://github.com/huggingface/datasets/pull/3175
[ "This looks great, thank you!", "Thanks !\r\n\r\nFor some reason the new GIF is 6MB, which is a bit heavy for an image on a website. The previous one was around 200KB though which is perfect. For a good experience we usually expect images to be less than 500KB - otherwise for users with poor connection it takes t...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3175", "html_url": "https://github.com/huggingface/datasets/pull/3175", "diff_url": "https://github.com/huggingface/datasets/pull/3175.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3175.patch", "merged_at": "2021-11-03T10:07...
3,175
true
Asserts replaced by exceptions (huggingface#3171)
I've replaced two asserts with their proper exceptions following the guidelines described in issue #3171 by following the contributing guidelines. PS: This is one of my first PRs, hoping I don't break anything!
https://github.com/huggingface/datasets/pull/3174
[ "Your first PR went smoothly, well done!\r\nYou are welcome to continue contributing to this project.\r\nGrΓ cies, @joseporiolayats! πŸ˜‰ " ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3174", "html_url": "https://github.com/huggingface/datasets/pull/3174", "diff_url": "https://github.com/huggingface/datasets/pull/3174.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3174.patch", "merged_at": "2021-10-29T13:08...
3,174
true
Fix issue with filelock filename being too long on encrypted filesystems
Infer max filename length in filelock on Unix-like systems. Should fix problems on encrypted filesystems such as eCryptfs. Fix #2924 cc: @lmmx
https://github.com/huggingface/datasets/pull/3173
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3173", "html_url": "https://github.com/huggingface/datasets/pull/3173", "diff_url": "https://github.com/huggingface/datasets/pull/3173.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3173.patch", "merged_at": "2021-10-29T09:42...
3,173
true
`SystemError 15` thrown in `Dataset.__del__` when using `Dataset.map()` with `num_proc>1`
## Describe the bug I use `datasets.map` to preprocess some data in my application. The error `SystemError 15` is thrown at the end of the execution of `Dataset.map()` (only with `num_proc>1`. Traceback included bellow. The exception is raised only when the code runs within a specific context. Despite ~10h spent ...
https://github.com/huggingface/datasets/issues/3172
[ "NB: even if the error is raised, the dataset is successfully cached. So restarting the script after every `map()` allows to ultimately run the whole preprocessing. But this prevents to realistically run the code over multiple nodes.", "Hi,\r\n\r\nIt's not easy to debug the problem without the script. I may be wr...
null
3,172
false
Raise exceptions instead of using assertions for control flow
Motivated by https://github.com/huggingface/transformers/issues/12789 in Transformers, one welcoming change would be replacing assertions with proper exceptions. The only type of assertions we should keep are those used as sanity checks. Currently, there is a total of 87 files with the `assert` statements (located u...
https://github.com/huggingface/datasets/issues/3171
[ "Adding the remaining tasks for this issue to help new code contributors. \r\n$ cd src/datasets && ack assert -lc \r\n- [x] commands/convert.py:1\r\n- [x] arrow_reader.py:3\r\n- [x] load.py:7\r\n- [x] utils/py_utils.py:2\r\n- [x] features/features.py:9\r\n- [x] arrow_writer.py:7\r\n- [x] search.py:6\r\n- [x] table...
null
3,171
false
Preserve ordering in `zip_dict`
Replace `set` with the `unique_values` generator in `zip_dict`. This PR fixes the problem with the different ordering of the example keys across different Python sessions caused by the `zip_dict` call in `Features.decode_example`.
https://github.com/huggingface/datasets/pull/3170
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3170", "html_url": "https://github.com/huggingface/datasets/pull/3170", "diff_url": "https://github.com/huggingface/datasets/pull/3170.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3170.patch", "merged_at": "2021-10-29T13:09...
3,170
true
Configurable max filename length in file locks
Resolve #2924 (https://github.com/huggingface/datasets/issues/2924#issuecomment-952330956) wherein the assumption of file lock maximum filename length to be 255 raises an OSError on encrypted drives (ecryptFS on Linux uses part of the lower filename, reducing the maximum filename size to 143). Allowing this limit to be...
https://github.com/huggingface/datasets/pull/3169
[ "I've also added environment variable configuration so that this can be configured once per machine (e.g. in a `.bashrc` file), as is already done for a few other config variables here.", "Cancelling PR in favour of @mariosasko's in #3173" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3169", "html_url": "https://github.com/huggingface/datasets/pull/3169", "diff_url": "https://github.com/huggingface/datasets/pull/3169.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3169.patch", "merged_at": null }
3,169
true
OpenSLR/83 is empty
## Describe the bug As the summary says, openslr / SLR83 / train is empty. The dataset returned after loading indicates there are **zero** rows. The correct number should be **17877**. ## Steps to reproduce the bug ```python import datasets datasets.load_dataset('openslr', 'SLR83') ``` ## Expected resul...
https://github.com/huggingface/datasets/issues/3168
[ "Hi @tyrius02, thanks for reporting. I see you self-assigned this issue: are you working on this?", "@albertvillanova Yes. Figured I introduced the broken config, I should fix it too.\r\n\r\nI've got it working, but I'm struggling with one of the tests. I've started a PR so I/we can work through it.", "Looks li...
null
3,168
false
bookcorpusopen no longer works
## Describe the bug When using the latest version of datasets (1.14.0), I cannot use the `bookcorpusopen` dataset. The process blocks always around `9924 examples [00:06, 1439.61 examples/s]` when preparing the dataset. I also noticed that after half an hour the process is automatically killed because of the RAM usa...
https://github.com/huggingface/datasets/issues/3167
[ "Hi ! Thanks for reporting :) I think #3280 should fix this", "I tried with the latest changes from #3280 on google colab and it worked fine :)\r\nWe'll do a new release soon, in the meantime you can use the updated version with:\r\n```python\r\nload_dataset(\"bookcorpusopen\", revision=\"master\")\r\n```", "Fi...
null
3,167
false
Deprecate prepare_module
In version 1.13, `prepare_module` was deprecated. This PR adds a deprecation warning and removes it from all the library, using `dataset_module_factory` or `metric_module_factory` instead. Fix #3165.
https://github.com/huggingface/datasets/pull/3166
[ "Sounds good, thanks !" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3166", "html_url": "https://github.com/huggingface/datasets/pull/3166", "diff_url": "https://github.com/huggingface/datasets/pull/3166.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3166.patch", "merged_at": "2021-11-05T09:27...
3,166
true
Deprecate prepare_module
In version 1.13, `prepare_module` was deprecated. Add deprecation warning and remove its usage from all the library.
https://github.com/huggingface/datasets/issues/3165
[]
null
3,165
false
Add raw data files to the Hub with GitHub LFS for canonical dataset
I'm interested in sharing the CaseHOLD dataset (https://arxiv.org/abs/2104.08671) as a canonical dataset on the HuggingFace Hub and would like to add the raw data files to the Hub with GitHub LFS, since it seems like a more sustainable long term storage solution, compared to other storage solutions available to my team...
https://github.com/huggingface/datasets/issues/3164
[ "Hi @zlucia, I would actually suggest hosting the dataset as a huggingface.co-hosted dataset.\r\n\r\nThe only difference with a \"canonical\"/legacy dataset is that it's nested under an organization (here `stanford` or `stanfordnlp` for instance – completely up to you) but then you can upload your data using git-lf...
null
3,164
false
Add Image feature
Adds the Image feature. This feature is heavily inspired by the recently added Audio feature (#2324). Currently, this PR is pretty simple. Some considerations that need further discussion: * I've decided to use `Pillow`/`PIL` as the image decoding library. Another candidate I considered is `torchvision`, mostly bec...
https://github.com/huggingface/datasets/pull/3163
[ "Awesome, looking forward to using it :)", "Few additional comments:\r\n* the current API doesn't meet the requirements mentioned in #3145 (e.g. image mime-type). However, this will be doable soon as we also plan to store image bytes alongside paths in arrow files (see https://github.com/huggingface/datasets/pull...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3163", "html_url": "https://github.com/huggingface/datasets/pull/3163", "diff_url": "https://github.com/huggingface/datasets/pull/3163.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3163.patch", "merged_at": "2021-12-06T17:49...
3,163
true
`datasets-cli test` should work with datasets without scripts
It would be really useful to be able to run `datasets-cli test`for datasets that don't have scripts attached to them (whether the datasets are private or not). I wasn't able to run the script for a private test dataset that I had created on the hub (https://huggingface.co/datasets/huggingface/DataMeasurementsTest/t...
https://github.com/huggingface/datasets/issues/3162
[ "> It would be really useful to be able to run `datasets-cli test`for datasets that don't have scripts attached to them (whether the datasets are private or not).\r\n> \r\n> I wasn't able to run the script for a private test dataset that I had created on the hub (https://huggingface.co/datasets/huggingface/DataMeas...
null
3,162
false
Add riddle_sense dataset
Adding a new dataset for QA with riddles. I'm confused about the tagging process because it looks like the streamlit app loads data from the current repo, so is it something that should be done after merging or off my fork?
https://github.com/huggingface/datasets/pull/3161
[ "@lhoestq \r\nI address all the comments, I think. Thanks! \r\n", "The five test fails are unrelated to this PR and fixed on master so we can ignore them" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3161", "html_url": "https://github.com/huggingface/datasets/pull/3161", "diff_url": "https://github.com/huggingface/datasets/pull/3161.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3161.patch", "merged_at": "2021-11-04T14:01...
3,161
true
Better error msg if `len(predictions)` doesn't match `len(references)` in metrics
Improve the error message in `Metric.add_batch` if `len(predictions)` doesn't match `len(references)`. cc: @BramVanroy (feel free to test this code on your examples and review this PR)
https://github.com/huggingface/datasets/pull/3160
[ "Can't test this now but it may be a good improvement indeed.", "I added a function, but it only works with the `list` type. For arrays/tensors, we delegate formatting to the frameworks. " ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3160", "html_url": "https://github.com/huggingface/datasets/pull/3160", "diff_url": "https://github.com/huggingface/datasets/pull/3160.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3160.patch", "merged_at": "2021-11-05T09:31...
3,160
true
Make inspect.get_dataset_config_names always return a non-empty list
Make all named configs cases, so that no special unnamed config case needs to be handled differently. Fix #3135.
https://github.com/huggingface/datasets/pull/3159
[ "This PR is already working (although not very beautiful; see below): the idea was to have the `DatasetModule.builder_kwargs` accessible from the `builder_cls`, so that this can generate the default builder config (at the class level, without requiring the builder to be instantiated).\r\n\r\nI have a plan for a fol...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3159", "html_url": "https://github.com/huggingface/datasets/pull/3159", "diff_url": "https://github.com/huggingface/datasets/pull/3159.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3159.patch", "merged_at": "2021-10-28T05:44...
3,159
true
Fix string encoding for Value type
Some metrics have `string` features but currently it fails if users pass integers instead. Indeed feature encoding that handles the conversion of the user's objects to the right python type is missing a case for `string`, while it already works as expected for integers, floats and booleans Here is an example code th...
https://github.com/huggingface/datasets/pull/3158
[ "That was fast! \r\n" ]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3158", "html_url": "https://github.com/huggingface/datasets/pull/3158", "diff_url": "https://github.com/huggingface/datasets/pull/3158.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3158.patch", "merged_at": "2021-10-25T14:12...
3,158
true
Fixed: duplicate parameter and missing parameter in docstring
changing duplicate parameter `data_files` in `DatasetBuilder.__init__` to the missing parameter `data_dir`
https://github.com/huggingface/datasets/pull/3157
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3157", "html_url": "https://github.com/huggingface/datasets/pull/3157", "diff_url": "https://github.com/huggingface/datasets/pull/3157.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3157.patch", "merged_at": "2021-10-25T14:02...
3,157
true
Illegal instruction (core dumped) at datasets import
## Describe the bug I install datasets using conda and when I import datasets I get: "Illegal instruction (core dumped)" ## Steps to reproduce the bug ``` conda create --prefix path/to/env conda activate path/to/env conda install -c huggingface -c conda-forge datasets # exits with output "Illegal instruction...
https://github.com/huggingface/datasets/issues/3155
[ "It seems to be an issue with how conda-forge is building the binaries. It works on some machines, but not a machine with AMD Opteron 8384 processors." ]
null
3,155
false
Sacrebleu unexpected behaviour/requirement for data format
## Describe the bug When comparing with the original `sacrebleu` implementation, the `datasets` implementation does some strange things that I do not quite understand. This issue was triggered when I was trying to implement TER and found the datasets implementation of BLEU [here](https://github.com/huggingface/dataset...
https://github.com/huggingface/datasets/issues/3154
[ "Hi @BramVanroy!\r\n\r\nGood question. This project relies on PyArrow (tables) to store data too big to fit in RAM. In the case of metrics, this means that the number of predictions and references has to match to form a table.\r\n\r\nThat's why your example throws an error even though it matches the schema:\r\n```p...
null
3,154
false
Add TER (as implemented in sacrebleu)
Implements TER (Translation Edit Rate) as per its implementation in sacrebleu. Sacrebleu for BLEU scores is already implemented in `datasets` so I thought this would be a nice addition. I started from the sacrebleu implementation, as the two metrics have a lot in common. Verified with sacrebleu's [testing suite](...
https://github.com/huggingface/datasets/pull/3153
[ "The problem appears to stem from the omission of the lines that you mentioned. If you add them back and try examples from [this](https://huggingface.co/docs/datasets/using_metrics.html) tutorial (sacrebleu metric example) the code you implemented works fine.\r\n\r\nI think the purpose of these lines is follows:\r\...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3153", "html_url": "https://github.com/huggingface/datasets/pull/3153", "diff_url": "https://github.com/huggingface/datasets/pull/3153.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3153.patch", "merged_at": "2021-11-02T11:04...
3,153
true
Fix some typos in the documentation
null
https://github.com/huggingface/datasets/pull/3152
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3152", "html_url": "https://github.com/huggingface/datasets/pull/3152", "diff_url": "https://github.com/huggingface/datasets/pull/3152.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3152.patch", "merged_at": "2021-10-25T14:03...
3,152
true
Re-add faiss to windows testing suite
In recent versions, `faiss-cpu` seems to be available for Windows as well. See the [PyPi page](https://pypi.org/project/faiss-cpu/#files) to confirm. We can therefore included it for Windows in the setup file. At first tests didn't pass due to problems with permissions as caused by `NamedTemporaryFile` on Windows. T...
https://github.com/huggingface/datasets/pull/3151
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3151", "html_url": "https://github.com/huggingface/datasets/pull/3151", "diff_url": "https://github.com/huggingface/datasets/pull/3151.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3151.patch", "merged_at": "2021-11-02T10:06...
3,151
true
Faiss _is_ available on Windows
In the setup file, I find the following: https://github.com/huggingface/datasets/blob/87c71b9c29a40958973004910f97e4892559dfed/setup.py#L171 However, FAISS does install perfectly fine on Windows on my system. You can also confirm this on the [PyPi page](https://pypi.org/project/faiss-cpu/#files), where Windows wh...
https://github.com/huggingface/datasets/issues/3150
[ "Sure, feel free to open a PR." ]
null
3,150
false
Add CMU Hinglish DoG Dataset for MT
Address part of #2841 Added the CMU Hinglish DoG Dataset as in GLUECoS. Added it as a seperate dataset as unlike other tasks of GLUE CoS this can't be evaluated for a BERT like model. Consists of parallel dataset between Hinglish (Hindi-English) and English, can be used for Machine Translation between the two. ...
https://github.com/huggingface/datasets/pull/3149
[ "Hi @lhoestq, thanks a lot for the help. I have moved the part as suggested. \r\nAlthough still while running the dummy data script, I face this issue\r\n\r\n```\r\nTraceback (most recent call last):\r\n File \"/home/ishan/anaconda3/bin/datasets-cli\", line 8, in <module>\r\n sys.exit(main())\r\n File \"/home/...
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3149", "html_url": "https://github.com/huggingface/datasets/pull/3149", "diff_url": "https://github.com/huggingface/datasets/pull/3149.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3149.patch", "merged_at": "2021-11-15T10:27...
3,149
true
Streaming with num_workers != 0
## Describe the bug When using dataset streaming with pytorch DataLoader, the setting num_workers to anything other than 0 causes the code to freeze forever before yielding the first batch. The code owner is likely @lhoestq ## Steps to reproduce the bug For your convenience, we've prepped a colab notebook th...
https://github.com/huggingface/datasets/issues/3148
[ "I can confirm that I was able to reproduce the bug. This seems odd given that #3423 reports duplicate data retrieval when `num_workers` and `streaming` are used together, which is obviously different from what is reported here. ", "Any update? A possible solution is to have multiple arrow files as shards, and ha...
null
3,148
false
Fix CLI test to ignore verfications when saving infos
Fix #3146.
https://github.com/huggingface/datasets/pull/3147
[]
{ "url": "https://api.github.com/repos/huggingface/datasets/pulls/3147", "html_url": "https://github.com/huggingface/datasets/pull/3147", "diff_url": "https://github.com/huggingface/datasets/pull/3147.diff", "patch_url": "https://github.com/huggingface/datasets/pull/3147.patch", "merged_at": "2021-10-27T08:01...
3,147
true
CLI test command throws NonMatchingSplitsSizesError when saving infos
When trying to generate a datset JSON metadata, a `NonMatchingSplitsSizesError` is thrown: ``` $ datasets-cli test datasets/arabic_billion_words --save_infos --all_configs Testing builder 'Alittihad' (1/10) Downloading and preparing dataset arabic_billion_words/Alittihad (download: 332.13 MiB, generated: Unknown si...
https://github.com/huggingface/datasets/issues/3146
[]
null
3,146
false