html_url stringlengths 48 51 | title stringlengths 5 268 | comments stringlengths 63 51.8k | body stringlengths 0 36.2k ⌀ | comment_length int64 16 1.52k | text stringlengths 164 54.1k | embeddings list |
|---|---|---|---|---|---|---|
https://github.com/huggingface/datasets/issues/1942 | [experiment] missing default_experiment-1-0.arrow | I just opened #1966 to fix this :)
@stas00 if have a chance feel free to try it ! | the original report was pretty bad and incomplete - my apologies!
Please see the complete version here: https://github.com/huggingface/datasets/issues/1942#issuecomment-786336481
------------
As mentioned here https://github.com/huggingface/datasets/issues/1939 metrics don't get cached, looking at my local `~/... | 19 | [experiment] missing default_experiment-1-0.arrow
the original report was pretty bad and incomplete - my apologies!
Please see the complete version here: https://github.com/huggingface/datasets/issues/1942#issuecomment-786336481
------------
As mentioned here https://github.com/huggingface/datasets/issues/19... | [
0.0502676554,
-0.0307855103,
0.058525674,
0.1668777615,
0.0344586708,
0.0303150676,
0.1830068678,
0.2939562201,
0.2358810008,
0.1017371789,
0.1023577899,
0.1887697279,
-0.3214571476,
0.0011923468,
0.0528262481,
-0.0287235118,
-0.0502644815,
-0.0096709365,
-0.358309865,
-0.12540... |
https://github.com/huggingface/datasets/issues/1941 | Loading of FAISS index fails for index_name = 'exact' | Works great 👍 I just put a minor comment on the commit, I think you meant to pass the `train_size` from the one obtained from the config.
Thanks for a quick response! | Hi,
It looks like loading of FAISS index now fails when using index_name = 'exact'.
For example, from the RAG [model card](https://huggingface.co/facebook/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).
Running `transformers==4.3.2` and datasets installed from source o... | 32 | Loading of FAISS index fails for index_name = 'exact'
Hi,
It looks like loading of FAISS index now fails when using index_name = 'exact'.
For example, from the RAG [model card](https://huggingface.co/facebook/rag-token-nq?fbclid=IwAR3bTfhls5U_t9DqsX2Vzb7NhtRHxJxfQ-uwFT7VuCPMZUM2AdAlKF_qkI8#usage).
Running `t... | [
-0.0889259577,
-0.1103836596,
0.0109982733,
0.0540016182,
0.410166651,
-0.0569177046,
0.2968125045,
0.2508974671,
0.2696112096,
0.1997553408,
-0.2217195779,
0.1398440152,
0.1363161802,
-0.1376477331,
-0.1762497127,
0.2843904495,
0.2129582018,
0.1230241507,
-0.0870197266,
-0.102... |
https://github.com/huggingface/datasets/issues/1940 | Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()` | Thanks for the report !
Currently we don't have a way to let the user easily disable this behavior.
However I agree that we should support stateful processing functions, ideally by removing `does_function_return_dict`.
We needed this function in order to know whether the `map` functions needs to write data or no... | Hi there!
In my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows/examples already selected per each class, which are the ones I want to keep in the end:
```python
... | 123 | Side effect when filtering data due to `does_function_return_dict` call in `Dataset.map()`
Hi there!
In my codebase I have a function to filter rows in a dataset, selecting only a certain number of examples per class. The function passes a extra argument to maintain a counter of the number of dataset rows/examples... | [
-0.4214116931,
0.0289576333,
-0.1575036347,
0.0184578039,
-0.0844257548,
-0.3201850951,
0.2051183581,
0.1946579218,
0.266033113,
0.1206626222,
0.2073838711,
0.5199098587,
-0.0788981766,
0.0967266113,
0.0301457271,
0.2262495458,
0.1668143868,
-0.0325006172,
-0.1896090806,
-0.123... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | Thanks for reporting and for all the details and suggestions.
I'm totally in favor of having a HF_DATASETS_OFFLINE env variable to disable manually all the connection checks, remove retries etc.
Moreover you may know that the use case that you are mentioning is already supported from `datasets` 1.3.0, i.e. you al... | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 156 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | I'm on master, so using all the available bells and whistles already.
If you look at the common issues - it for example tries to look up files if they appear in `_PACKAGED_DATASETS_MODULES` which it shouldn't do.
--------------
Yes, there is a nuance to it. As I mentioned it's firewalled - that is it has a net... | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 257 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | Ok understand better the hanging issue.
I guess catching connection errors is not enough, we should also avoid all the hangings.
Currently the offline mode tests are only done by simulating an instant connection fail that returns an error, let's have another connection mock that hangs instead.
I'll also take a loo... | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 61 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | FWIW, I think instant failure on the behalf of a network call is the simplest solution to correctly represent the environment and having the caller to sort it out is the next thing to do, since here it is the case of having no functional network, it's just that the software doesn't know this is the case, because there ... | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 88 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | Update on this:
I managed to create a mock environment for tests that makes the connections hang until timeout.
I managed to reproduce the issue you're having in this environment.
I'll update the offline test cases to also test the robustness to connection hangings, and make sure we set proper timeouts where it... | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 65 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1939 | [firewalled env] OFFLINE mode | I lost access to the firewalled setup, but I emulated it with:
```
sudo ufw enable
sudo ufw default deny outgoing
```
(thanks @mfuntowicz)
I was able to test `HF_DATASETS_OFFLINE=1` and it worked great - i.e. didn't try to reach out with it and used the cached files instead.
Thank you! | This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a sample program. There are 2 pos... | 51 | [firewalled env] OFFLINE mode
This issue comes from a need to be able to run `datasets` in a firewalled env, which currently makes the software hang until it times out, as it's unable to complete the network calls.
I propose the following approach to solving this problem, using the example of `run_seq2seq.py` as a... | [
-0.4748800993,
0.1685328484,
-0.0625534207,
0.1050562114,
0.044311218,
-0.2284920216,
0.4647971094,
0.0716797411,
0.2545912862,
0.0485305525,
0.0184395593,
0.0636147857,
0.0845068023,
0.1993041337,
0.0180915035,
0.1079675481,
0.0199453812,
-0.10963884,
-0.2528403103,
0.06371635... |
https://github.com/huggingface/datasets/issues/1924 | Anonymous Dataset Addition (i.e Anonymous PR?) | Hi !
I guess you can add a dataset without the fields that must be kept anonymous, and then update those when the anonymity period is over.
You can also make the PR from an anonymous org.
Pinging @yjernite just to make sure it's ok | Hello,
Thanks a lot for your librairy.
We plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ?
Cheers
@eusip | 45 | Anonymous Dataset Addition (i.e Anonymous PR?)
Hello,
Thanks a lot for your librairy.
We plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ?
Cheers
@eusip
Hi !
I guess you can add a dataset without th... | [
-0.2978557944,
0.5536883473,
-0.0238175262,
-0.1903210729,
-0.0758948326,
-0.1079590023,
0.5019997358,
0.0080971094,
0.0642960221,
0.0534696281,
0.0712102056,
0.1038330495,
0.0041554756,
0.1634859145,
-0.0382041335,
0.0795786679,
0.0454186834,
-0.0966099054,
0.0962191373,
-0.20... |
https://github.com/huggingface/datasets/issues/1924 | Anonymous Dataset Addition (i.e Anonymous PR?) | Hello,
I would prefer to do the reverse: adding a link to an anonymous paper without the people names/institution in the PR. Would it be conceivable ?
Cheers
| Hello,
Thanks a lot for your librairy.
We plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ?
Cheers
@eusip | 28 | Anonymous Dataset Addition (i.e Anonymous PR?)
Hello,
Thanks a lot for your librairy.
We plan to submit a paper on OpenReview using the Anonymous setting. Is it possible to add a new dataset without breaking the anonimity, with a link to the paper ?
Cheers
@eusip
Hello,
I would prefer to do the reverse: addi... | [
-0.2011671811,
0.4423333704,
0.0201970674,
-0.0765714347,
-0.0947152153,
-0.0944124088,
0.438736558,
-0.0027228699,
0.0362779014,
0.0934397057,
-0.0507458821,
0.0669826567,
0.003766044,
0.0874345526,
0.1370115578,
0.1177667826,
0.1308903992,
-0.0895972922,
0.209927842,
-0.24951... |
https://github.com/huggingface/datasets/issues/1922 | How to update the "wino_bias" dataset | Hi @JieyuZhao !
You can edit the dataset card of wino_bias to update the URL via a Pull Request. This would be really appreciated :)
The dataset card is the README.md file you can find at https://github.com/huggingface/datasets/tree/master/datasets/wino_bias
Also the homepage url is also mentioned in the wino_bi... | Hi all,
Thanks for the efforts to collect all the datasets! But I think there is a problem with the wino_bias dataset. The current link is not correct. How can I update that?
Thanks! | 89 | How to update the "wino_bias" dataset
Hi all,
Thanks for the efforts to collect all the datasets! But I think there is a problem with the wino_bias dataset. The current link is not correct. How can I update that?
Thanks!
Hi @JieyuZhao !
You can edit the dataset card of wino_bias to update the URL via a Pul... | [
-0.3595492244,
0.1663749367,
-0.0910966992,
0.1053338125,
0.0010984784,
0.1791955084,
-0.0393257327,
0.0637113377,
0.0622652397,
-0.1287821382,
-0.2621753812,
-0.0001707154,
0.257883817,
0.1593093723,
0.2776814401,
-0.3877266943,
0.1113262996,
0.1100791171,
0.1167295799,
-0.211... |
https://github.com/huggingface/datasets/issues/1919 | Failure to save with save_to_disk | Hi thanks for reporting and for proposing a fix :)
I just merged a fix, feel free to try it from the master branch ! | When I try to save a dataset locally using the `save_to_disk` method I get the error:
```bash
FileNotFoundError: [Errno 2] No such file or directory: '/content/squad/train/squad-train.arrow'
```
To replicate:
1. Install `datasets` from master
2. Run this code:
```python
from datasets import load... | 25 | Failure to save with save_to_disk
When I try to save a dataset locally using the `save_to_disk` method I get the error:
```bash
FileNotFoundError: [Errno 2] No such file or directory: '/content/squad/train/squad-train.arrow'
```
To replicate:
1. Install `datasets` from master
2. Run this code:
```p... | [
-0.0436869413,
0.1429487616,
0.0570115484,
0.1895617694,
0.5194175243,
0.2885243595,
0.1922132075,
0.2388130426,
-0.1313644052,
0.1859048009,
0.1415780485,
0.368059665,
-0.3196104467,
-0.250821054,
0.069875069,
0.04740569,
0.2884761691,
0.0203922167,
-0.0373371318,
-0.082811527... |
https://github.com/huggingface/datasets/issues/1915 | Unable to download `wiki_dpr` | Thanks for reporting ! This is a bug. For now feel free to set `ignore_verifications=False` in `load_dataset`.
I'm working on a fix | I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings/no index. In order to do so, I ran:
`curr_dataset = load_dataset("wiki_dpr", embeddings_name="multiset", index_name="no_index")`
However, I got the following error:
`datasets.utils.i... | 22 | Unable to download `wiki_dpr`
I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings/no index. In order to do so, I ran:
`curr_dataset = load_dataset("wiki_dpr", embeddings_name="multiset", index_name="no_index")`
However, I got the fol... | [
-0.359107852,
-0.3630982935,
-0.0504425764,
0.3148261607,
0.3717035949,
0.4270488024,
0.3198952079,
0.0625340641,
0.238521561,
0.1804991513,
0.1647018641,
-0.0206089485,
0.0954112783,
-0.0210909862,
-0.1673089117,
-0.0334006473,
0.1001308858,
0.1070559844,
0.3710588217,
-0.0622... |
https://github.com/huggingface/datasets/issues/1915 | Unable to download `wiki_dpr` | I just merged a fix :)
We'll do a patch release soon. In the meantime feel free to try it from the master branch
Thanks again for reporting ! | I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings/no index. In order to do so, I ran:
`curr_dataset = load_dataset("wiki_dpr", embeddings_name="multiset", index_name="no_index")`
However, I got the following error:
`datasets.utils.i... | 29 | Unable to download `wiki_dpr`
I am trying to download the `wiki_dpr` dataset. Specifically, I want to download `psgs_w100.multiset.no_index` with no embeddings/no index. In order to do so, I ran:
`curr_dataset = load_dataset("wiki_dpr", embeddings_name="multiset", index_name="no_index")`
However, I got the fol... | [
-0.359107852,
-0.3630982935,
-0.0504425764,
0.3148261607,
0.3717035949,
0.4270488024,
0.3198952079,
0.0625340641,
0.238521561,
0.1804991513,
0.1647018641,
-0.0206089485,
0.0954112783,
-0.0210909862,
-0.1673089117,
-0.0334006473,
0.1001308858,
0.1070559844,
0.3710588217,
-0.0622... |
https://github.com/huggingface/datasets/issues/1911 | Saving processed dataset running infinitely | am suspicious of this thing? what's the purpose of this? pickling and unplickling
`self = pickle.loads(pickle.dumps(self))`
```
def save_to_disk(self, dataset_path: str, fs=None):
"""
Saves a dataset to a dataset directory, or in a filesystem using either :class:`datasets.filesystem.S3FileSys... | I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so I used a hack to use pyarrow filter table func... | 103 | Saving processed dataset running infinitely
I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so ... | [
-0.2176190019,
0.1967645884,
-0.1150818169,
0.2636954486,
0.1485698372,
-0.07311932,
0.1913731247,
0.2333744764,
-0.1932331622,
-0.0838295892,
0.0687727109,
0.3565645516,
-0.1087137982,
0.2635344267,
0.0180144235,
0.2948513627,
0.2866331637,
0.1167816892,
0.2218980491,
0.044339... |
https://github.com/huggingface/datasets/issues/1911 | Saving processed dataset running infinitely | Tried finding the root cause but was unsuccessful.
I am using lazy tokenization with `dataset.set_transform()`, it works like a charm with almost same performance as pre-compute. | I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so I used a hack to use pyarrow filter table func... | 26 | Saving processed dataset running infinitely
I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so ... | [
-0.3561965227,
0.1683891267,
-0.0956781805,
0.2705767751,
0.2809064388,
-0.0555667207,
0.2274561375,
0.2229017764,
-0.3108583689,
-0.1395981759,
0.1082136929,
0.175986439,
-0.1438533813,
0.2344980687,
0.0023733405,
0.2562197745,
0.3514586091,
0.1079184264,
0.3704065681,
0.10870... |
https://github.com/huggingface/datasets/issues/1911 | Saving processed dataset running infinitely | Hi ! This very probably comes from the hack you used.
The pickling line was added an a sanity check because save_to_disk uses the same assumptions as pickling for a dataset object. The main assumption is that memory mapped pyarrow tables must be reloadable from the disk. In your case it's not possible since you alte... | I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so I used a hack to use pyarrow filter table func... | 191 | Saving processed dataset running infinitely
I have a text dataset of size 220M.
For pre-processing, I need to tokenize this and filter rows with the large sequence.
My tokenization took roughly 3hrs. I used map() with batch size 1024 and multi-process with 96 processes.
filter() function was way to slow, so ... | [
-0.3835665286,
0.2207913399,
-0.0753483772,
0.2524214983,
0.1676333398,
-0.0859724581,
0.0995909274,
0.2425321341,
-0.1649822444,
-0.0584912151,
0.033023417,
0.4020829499,
-0.0848711133,
0.2319576293,
0.0172860175,
0.2511723042,
0.1903819144,
0.0988865197,
0.1988591105,
0.19222... |
https://github.com/huggingface/datasets/issues/1907 | DBPedia14 Dataset Checksum bug? | Hi ! :)
This looks like the same issue as https://github.com/huggingface/datasets/issues/1856
Basically google drive has quota issues that makes it inconvenient for downloading files.
If the quota of a file is exceeded, you have to wait 24h for the quota to reset (which is painful).
The error says that the c... | Hi there!!!
I've been using successfully the DBPedia dataset (https://huggingface.co/datasets/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:
```
Traceback (most recent call last):
File "./conditional_classification/basic_pipeline.py", line 178, i... | 75 | DBPedia14 Dataset Checksum bug?
Hi there!!!
I've been using successfully the DBPedia dataset (https://huggingface.co/datasets/dbpedia_14) with my codebase in the last couple of weeks, but in the last couple of days now I get this error:
```
Traceback (most recent call last):
File "./conditional_classificati... | [
-0.2483559996,
0.3508054316,
-0.1227643937,
0.262452513,
0.1046653613,
-0.0048448113,
0.3205701411,
0.5194302797,
-0.0877417848,
-0.0413589142,
0.0776256472,
-0.1982264072,
0.0031295752,
0.3474116027,
-0.2444945425,
-0.0099744089,
0.0986698866,
-0.1543820947,
-0.0066187195,
0.0... |
https://github.com/huggingface/datasets/issues/1906 | Feature Request: Support for Pandas `Categorical` | We already have a ClassLabel type that does this kind of mapping between the label ids (integers) and actual label values (strings).
I wonder if actually we should use the DictionaryType from Arrow and the Categorical type from pandas for the `datasets` ClassLabel feature type.
Currently ClassLabel corresponds to `... | ```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_... | 69 | Feature Request: Support for Pandas `Categorical`
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to ... | [
0.0230096038,
-0.010273939,
-0.1426059604,
0.2076645494,
0.2370132059,
0.1672787219,
0.1267406344,
0.2487183362,
-0.0738423392,
-0.2155076712,
0.1673995703,
0.2824070454,
-0.2035115957,
0.4252625406,
0.13274616,
-0.2236517221,
0.1246827915,
0.0284454785,
0.0778504834,
0.3117168... |
https://github.com/huggingface/datasets/issues/1906 | Feature Request: Support for Pandas `Categorical` | Now that I've heard you explain ClassLabel, that makes a lot of sense! While DictionaryType for Arrow (I think) can have arbitrarily typed keys, so it won't cover all potential cases, pandas' Category is *probably* the most common use for that pyarrow type, and ClassLabel should match that perfectly?
Other thoughts... | ```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_... | 319 | Feature Request: Support for Pandas `Categorical`
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to ... | [
0.0230096038,
-0.010273939,
-0.1426059604,
0.2076645494,
0.2370132059,
0.1672787219,
0.1267406344,
0.2487183362,
-0.0738423392,
-0.2155076712,
0.1673995703,
0.2824070454,
-0.2035115957,
0.4252625406,
0.13274616,
-0.2236517221,
0.1246827915,
0.0284454785,
0.0778504834,
0.3117168... |
https://github.com/huggingface/datasets/issues/1906 | Feature Request: Support for Pandas `Categorical` | I think we can first keep the int64 precision but with an arrow Dictionary for ClassLabel, and focus on the connection with arrow and pandas.
In this scope, I really like the idea of checking for the dictionary type:
> Another idea that just occurred to me: add a branch in here to generate a ClassLabel if the dic... | ```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to the dictionary as well (for labels). I.e. to the py_... | 260 | Feature Request: Support for Pandas `Categorical`
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.Series(["a", "b", "c", "a"], dtype="category"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws NotImplementedError
# TODO(thom) this will need access to ... | [
0.0230096038,
-0.010273939,
-0.1426059604,
0.2076645494,
0.2370132059,
0.1672787219,
0.1267406344,
0.2487183362,
-0.0738423392,
-0.2155076712,
0.1673995703,
0.2824070454,
-0.2035115957,
0.4252625406,
0.13274616,
-0.2236517221,
0.1246827915,
0.0284454785,
0.0778504834,
0.3117168... |
https://github.com/huggingface/datasets/issues/1898 | ALT dataset has repeating instances in all splits | I just merged a fix, we'll do a patch release soon. Thanks again for reporting, and sorry for the inconvenience.
In the meantime you can load `ALT` using `datasets` from the master branch | The [ALT](https://huggingface.co/datasets/alt) dataset has all the same instances within each split :/
Seemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.
Would be great if this could be fixed :)
Added a snapshot of the contents from `exp... | 33 | ALT dataset has repeating instances in all splits
The [ALT](https://huggingface.co/datasets/alt) dataset has all the same instances within each split :/
Seemed like a great dataset for some experiments I wanted to carry out, especially since its medium-sized, and has all splits.
Would be great if this could be fi... | [
-0.2530450225,
-0.7312532663,
-0.0555955581,
0.4320777059,
0.317453891,
-0.173677057,
0.3799057305,
0.1424697191,
0.2794246376,
0.331989646,
-0.1614304185,
0.0980525017,
0.0777060986,
-0.0084827151,
-0.0531923808,
-0.0700328946,
-0.0420137085,
0.2138654143,
-0.0015913452,
-0.22... |
https://github.com/huggingface/datasets/issues/1895 | Bug Report: timestamp[ns] not recognized | Thanks for reporting !
You're right, `string_to_arrow` should be able to take `"timestamp[ns]"` as input and return the right pyarrow timestamp type.
Feel free to suggest a fix for `string_to_arrow` and open a PR if you want to contribute ! This would be very appreciated :)
To give you more context:
As you ma... | Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.
```
The fact... | 159 | Bug Report: timestamp[ns] not recognized
Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems ... | [
-0.1527937204,
0.3287225068,
0.0040783715,
0.087167725,
0.0947733223,
-0.0773304254,
0.4513311684,
0.3623487353,
-0.4917005301,
-0.2842442691,
0.2675410509,
0.5988476872,
-0.1803750843,
0.0181876477,
0.029280493,
-0.1218027398,
0.1887533218,
0.2048998326,
0.0169885419,
0.292965... |
https://github.com/huggingface/datasets/issues/1895 | Bug Report: timestamp[ns] not recognized | Thanks for the clarification @lhoestq !
This may be a little bit of a stupid question, but I wanted to clarify one more thing before I took a stab at this:
When the features get inferred, I believe they already have a pyarrow schema (https://github.com/huggingface/datasets/blob/master/src/datasets/arrow_dataset.p... | Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.
```
The fact... | 100 | Bug Report: timestamp[ns] not recognized
Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems ... | [
-0.0739535019,
0.2633907795,
0.023429187,
0.1172330901,
0.0697289258,
-0.1028006598,
0.34825176,
0.2648757398,
-0.5368142724,
-0.2700382471,
0.239961639,
0.4913218915,
-0.1497713029,
-0.0816279352,
0.0596877411,
-0.122853756,
0.1500018239,
0.1810296327,
0.0497961901,
0.27606067... |
https://github.com/huggingface/datasets/issues/1895 | Bug Report: timestamp[ns] not recognized | The objective in terms of design is to make it easy to create Features in a pythonic way. So for example we use a string to define a Value type.
That's why when inferring the Features from an arrow schema we have to find the right string definitions for Value types. I guess we could also have a constructor `Value.from... | Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.
```
The fact... | 116 | Bug Report: timestamp[ns] not recognized
Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems ... | [
-0.1146750525,
0.1922074407,
0.0449614413,
0.1444820166,
-0.0285736155,
-0.1225974411,
0.444470942,
0.2822775543,
-0.5322681069,
-0.3254839182,
0.3281250894,
0.5652284026,
-0.20688425,
0.088383399,
0.060448911,
-0.1087238789,
0.1322405934,
0.1888607442,
0.0694060698,
0.30265009... |
https://github.com/huggingface/datasets/issues/1895 | Bug Report: timestamp[ns] not recognized | OK I think I understand now:
Features are datasets' internal representation of a schema type, distinct from pyarrow's schema.
Value() corresponds to pyarrow's "primitive" types (e.g. `int` or `string`, but not things like `list` or `dict`).
`get_nested_type()` (https://github.com/huggingface/datasets/blob/master/s... | Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems to be a pyarrow data type.
```
The fact... | 96 | Bug Report: timestamp[ns] not recognized
Repro:
```
from datasets import Dataset
import pandas as pd
import pyarrow
df = pd.DataFrame(pd.date_range("2018-01-01", periods=3, freq="H"))
pyarrow.Table.from_pandas(df)
Dataset.from_pandas(df)
# Throws ValueError: Neither timestamp[ns] nor timestamp[ns]_ seems ... | [
-0.1425738186,
0.3246958256,
-0.0001188215,
0.1199890226,
0.0674352571,
-0.0765274167,
0.3947215676,
0.277019769,
-0.4754166603,
-0.2832739353,
0.2034533322,
0.5407938361,
-0.1618391126,
0.0006554835,
0.0657186732,
-0.103197813,
0.1609957963,
0.2012025118,
0.074615255,
0.296737... |
https://github.com/huggingface/datasets/issues/1894 | benchmarking against MMapIndexedDataset | Hi sam !
Indeed we can expect the performances to be very close since both MMapIndexedDataset and the `datasets` implem use memory mapping. With memory mapping what determines the I/O performance is the speed of your hard drive/SSD.
In terms of performance we're pretty close to the optimal speed for reading text, e... | I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB o... | 141 | benchmarking against MMapIndexedDataset
I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory an... | [
-0.3810418844,
-0.0305888429,
-0.1813575774,
0.2504055798,
-0.2935996354,
-0.0467079431,
0.0398224406,
0.1909919977,
-0.1505499333,
-0.0903934166,
-0.1634377688,
0.3356112242,
0.031131193,
-0.4226372242,
0.2201005071,
-0.1310539991,
0.3650218248,
0.0921235755,
-0.2799273431,
0.... |
https://github.com/huggingface/datasets/issues/1894 | benchmarking against MMapIndexedDataset | Also I would be interested to know what data types `MMapIndexedDataset` supports. Is there some documentation somewhere ? | I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB o... | 18 | benchmarking against MMapIndexedDataset
I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory an... | [
-0.2623204887,
-0.0587806515,
-0.1896095127,
0.2777882814,
-0.227725938,
-0.0737620369,
0.0111631993,
0.1357611269,
-0.2933742106,
-0.1568025649,
-0.1532004178,
0.3774216771,
0.0027403706,
-0.3709001839,
0.2234068811,
-0.1318498552,
0.3597462773,
0.1027275771,
-0.2230797857,
0.... |
https://github.com/huggingface/datasets/issues/1894 | benchmarking against MMapIndexedDataset | no docs haha, it's written to support integer numpy arrays.
You can build one in fairseq with, roughly:
```bash
wget https://s3.amazonaws.com/research.metamind.io/wikitext/wikitext-103-raw-v1.zip
unzip wikitext-103-raw-v1.zip
export dd=$HOME/fairseq-py/wikitext-103-raw
export mm_dir=$HOME/mmap_wikitext2
mk... | I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory and runs 1% slower for `wikitext103` (~1GB o... | 249 | benchmarking against MMapIndexedDataset
I am trying to benchmark my datasets based implementation against fairseq's [`MMapIndexedDataset`](https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L365) and finding that, according to psrecord, my `datasets` implem uses about 3% more CPU memory an... | [
-0.2432022691,
0.0233527608,
-0.1691764742,
0.1820508242,
-0.1022189632,
-0.0226157606,
0.1294503659,
0.2604652047,
-0.2088270187,
-0.0699170753,
-0.1556989551,
0.503326118,
-0.070250079,
-0.4188560545,
0.2817452848,
-0.0300164688,
0.3354426622,
0.1515544802,
-0.1571369916,
0.0... |
https://github.com/huggingface/datasets/issues/1893 | wmt19 is broken | This was also mentioned in https://github.com/huggingface/datasets/issues/488
The bucket where is data was stored seems to be unavailable now. Maybe we can change the URL to the ones in https://conferences.unite.un.org/uncorpus/en/downloadoverview ? | 1. Check which lang pairs we have: `--dataset_name wmt19`:
Please pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']
2. OK, let's pick `ru-en`:
`--dataset_name wmt19 --dataset_config "ru-en"`
no cookies:
```
Traceback (most recent c... | 30 | wmt19 is broken
1. Check which lang pairs we have: `--dataset_name wmt19`:
Please pick one among the available configs: ['cs-en', 'de-en', 'fi-en', 'gu-en', 'kk-en', 'lt-en', 'ru-en', 'zh-en', 'fr-de']
2. OK, let's pick `ru-en`:
`--dataset_name wmt19 --dataset_config "ru-en"`
no cookies:
```
Traceb... | [
-0.3531679809,
-0.4596048594,
-0.0412609763,
0.3225633502,
0.1783248633,
-0.0362051018,
0.1937318295,
0.23352696,
0.0660969839,
0.0606785864,
0.0147965243,
0.109032467,
-0.1255615205,
0.5707480311,
0.1197271943,
-0.0096023111,
0.0067276941,
0.0419427566,
-0.4998805523,
0.068993... |
https://github.com/huggingface/datasets/issues/1892 | request to mirror wmt datasets, as they are really slow to download | Yes that would be awesome. Not only the download speeds are awful, but also some files are missing.
We list all the URLs in the datasets/wmt19/wmt_utils.py so we can make a script to download them all and host on S3.
Also I think most of the materials are under the CC BY-NC-SA 3.0 license (must double check) so it sh... | Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you! | 78 | request to mirror wmt datasets, as they are really slow to download
Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you!
Yes that would be awesome. Not onl... | [
-0.1942381859,
-0.3241638541,
0.0704937354,
0.2124049962,
0.0100603672,
0.0652930588,
0.1663088799,
0.4544948936,
0.0562023148,
-0.1142763272,
-0.2518415749,
-0.073851712,
-0.0602756552,
0.3107830286,
-0.0818375871,
-0.0826896951,
-0.0564789549,
0.1170890778,
-0.6240909696,
-0.... |
https://github.com/huggingface/datasets/issues/1892 | request to mirror wmt datasets, as they are really slow to download | Yeah, the scripts are pretty ugly! A big refactor would make sense here...and I also remember that the datasets were veeery slow to download | Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you! | 24 | request to mirror wmt datasets, as they are really slow to download
Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you!
Yeah, the scripts are pretty ugly!... | [
-0.3805082738,
-0.175354138,
0.0075875446,
0.1037698463,
-0.0104691628,
0.05997774,
0.1670316011,
0.507707417,
0.2716600597,
-0.0307256971,
-0.2506196797,
-0.2216314822,
0.0339968018,
0.243268609,
-0.1237286851,
-0.2390305251,
-0.174991414,
0.1651735008,
-0.7791349888,
-0.17991... |
https://github.com/huggingface/datasets/issues/1892 | request to mirror wmt datasets, as they are really slow to download | I'm downloading them.
I'm starting with the ones hosted on http://data.statmt.org which are the slowest ones | Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you! | 16 | request to mirror wmt datasets, as they are really slow to download
Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you!
I'm downloading them.
I'm startin... | [
-0.4048578441,
-0.3140090406,
0.0369966552,
0.1695625484,
-0.013983977,
0.1483641118,
0.0976711735,
0.440671593,
0.2012497634,
-0.0260527022,
-0.3467168808,
-0.2557105124,
0.089427121,
0.1072441414,
-0.160569787,
-0.2548779249,
-0.0824238062,
0.1284086406,
-0.6802175045,
-0.137... |
https://github.com/huggingface/datasets/issues/1892 | request to mirror wmt datasets, as they are really slow to download | @lhoestq better to use our new git-based system than just raw S3, no? (that way we have built-in CDN etc.) | Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you! | 20 | request to mirror wmt datasets, as they are really slow to download
Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you!
@lhoestq better to use our new git... | [
-0.2889100611,
-0.3306779563,
0.0120120794,
0.109323144,
0.0525673963,
-0.0814258531,
0.1615228057,
0.4271323681,
0.0938708708,
-0.0190337282,
-0.3289591074,
-0.0868707746,
0.0092278542,
0.3145241141,
-0.2300494164,
-0.1343408376,
0.0110248467,
0.1879598349,
-0.5668649077,
-0.1... |
https://github.com/huggingface/datasets/issues/1892 | request to mirror wmt datasets, as they are really slow to download | Hi there! What about mirroring other datasets like [CCAligned](http://www.statmt.org/cc-aligned/) as well? All of them are really slow to download... | Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you! | 19 | request to mirror wmt datasets, as they are really slow to download
Would it be possible to mirror the wmt data files under hf? Some of them take hours to download and not because of the local speed. They are all quite small datasets, just extremely slow to download.
Thank you!
Hi there! What about mirroring oth... | [
-0.4533649385,
-0.2987291217,
0.02584018,
0.0953520015,
0.063311778,
0.1306117177,
0.1700921059,
0.3957925439,
0.1734650135,
-0.0313709266,
-0.3652692735,
-0.2737769783,
0.0889613926,
0.1364366263,
-0.2166112065,
-0.2842340767,
-0.0774133131,
0.1845021844,
-0.6560146809,
-0.094... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | I started working on this. My idea is to first add the pyarrow Table wrappers InMemoryTable and MemoryMappedTable that both implement what's necessary regarding copy/pickle. Then have another wrapper that takes the concatenation of InMemoryTable/MemoryMappedTable objects.
What's important here is that concatenating ... | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 55 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.3632806242,
0.1712489575,
-0.016625721,
0.1944020987,
0.099194631,
0.0310435034,
-0.2240343541,
0.2879286408,
-0.2779929042,
0.1697993726,
-0.0141433608,
0.6244733334,
-0.0076514403,
0.3612301946,
0.3155478537,
0.0208594948,
0.1660714746,
0.3219620287,
-0.4445088804,
0.21201... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | Hi @lhoestq @albertvillanova,
I checked the linked issues and PR, this seems like a great idea. Would you mind elaborating on the in-memory and memory-mapped datasets?
Based on my understanding, it is something like this, please correct me if I am wrong:
1. For in-memory datasets, we don't have any dataset files ... | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 129 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.3151937425,
0.0239814427,
-0.0091472091,
0.3324131072,
-0.0929978788,
-0.0160014015,
-0.1415652484,
0.0513882488,
-0.1234133914,
-0.0343253911,
0.0126751512,
0.5089059472,
0.0155091174,
0.4306792915,
0.1224203929,
0.0571519993,
0.1958597451,
0.1693576574,
-0.4769209027,
0.10... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | Hi ! Yes you're totally right about your two points :)
And in the case of a concatenated dataset, then we should reload each sub-table depending on whether it's in-memory or memory mapped. That means the dataset will be made of several blocks in order to keep track of what's from memory and what's memory mapped. Thi... | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 62 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.3308894336,
0.0809662193,
-0.0875296667,
0.1288807392,
0.1037874222,
0.021383414,
-0.0953436419,
0.2642481625,
-0.0994138345,
0.1963124722,
-0.0066627087,
0.5743990541,
-0.0273895394,
0.4745154083,
0.1390601844,
0.019964464,
0.1339084506,
0.270672977,
-0.4599793851,
0.099557... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | Hi @lhoestq
Thanks, that sounds nice. Can you explain where the issue of the double memory may arise? Also, why is the existing `concatenate_datasets` not sufficient for this purpose? | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 29 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.3921350241,
0.1095159948,
-0.0102483183,
0.3699242175,
0.0255512148,
0.1648332626,
-0.2444259375,
0.1994280517,
-0.1821033955,
0.1366394013,
0.049568139,
0.4666679204,
0.0200921427,
0.3617292941,
0.1092142239,
-0.0306007061,
0.1604873389,
0.2564074695,
-0.4017225206,
0.16292... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | Hi @lhoestq,
Will the `add_item` feature also help with lazy writing (or no caching) during `map`/`filter`? | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 16 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.4765585959,
0.0608364642,
-0.0707228556,
0.0521822609,
0.0756110325,
0.0150534166,
-0.2226112932,
0.3435709774,
0.0733317584,
0.0877744183,
0.029772846,
0.6137645245,
-0.0133162746,
0.4729112685,
0.115610823,
0.021717716,
0.1311556995,
0.2873374224,
-0.4156749845,
0.15891422... |
https://github.com/huggingface/datasets/issues/1877 | Allow concatenation of both in-memory and on-disk datasets | > Can you explain where the issue of the double memory may arise?
We have to keep each block (in-memory vs memory mapped) separated in order to be able to reload them with pickle.
On the other hand we also need to have the full table from mixed in-memory and memory mapped data in order to iterate or extract data co... | This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using the dataset._data_files).
This assumption is used for pickl... | 188 | Allow concatenation of both in-memory and on-disk datasets
This is a prerequisite for the addition of the `add_item` feature (see #1870).
Currently there is one assumption that we would need to change: a dataset is either fully in memory (dataset._data_files is empty), or the dataset can be reloaded from disk (using... | [
-0.4205480814,
0.1177791804,
-0.0135110943,
0.2829381824,
0.0721301436,
0.0939802006,
-0.1494366825,
0.2320159674,
-0.1620394439,
0.1522654593,
0.0206330474,
0.5102048516,
-0.0248011425,
0.357752353,
0.1537400782,
-0.0333431587,
0.1533058137,
0.2407621443,
-0.4325168133,
0.2023... |
https://github.com/huggingface/datasets/issues/1876 | load_dataset("multi_woz_v22") NonMatchingChecksumError | Thanks for reporting !
This is due to the changes made in the data files in the multiwoz repo: https://github.com/budzianowski/multiwoz/pull/59
I'm opening a PR to update the checksums of the data files. | Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumError(error_msg + str(bad_urls))
datasets.utils.info_utils.N... | 32 | load_dataset("multi_woz_v22") NonMatchingChecksumError
Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumErr... | [
-0.2244368196,
0.1764704585,
-0.0287197083,
0.1495724618,
0.1895724684,
0.0018985717,
0.3650580049,
0.4820297062,
0.2514896691,
0.166606307,
-0.0970347971,
0.1781591028,
-0.0822444707,
0.0992017165,
-0.184548825,
0.2732409537,
0.0765194371,
-0.0256412365,
-0.1682781428,
-0.0625... |
https://github.com/huggingface/datasets/issues/1876 | load_dataset("multi_woz_v22") NonMatchingChecksumError | I just merged the fix. It will be available in the new release of `datasets` later today.
You'll be able to get the new version with
```
pip install --upgrade datasets
``` | Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumError(error_msg + str(bad_urls))
datasets.utils.info_utils.N... | 32 | load_dataset("multi_woz_v22") NonMatchingChecksumError
Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumErr... | [
-0.2244368196,
0.1764704585,
-0.0287197083,
0.1495724618,
0.1895724684,
0.0018985717,
0.3650580049,
0.4820297062,
0.2514896691,
0.166606307,
-0.0970347971,
0.1781591028,
-0.0822444707,
0.0992017165,
-0.184548825,
0.2732409537,
0.0765194371,
-0.0256412365,
-0.1682781428,
-0.0625... |
https://github.com/huggingface/datasets/issues/1876 | load_dataset("multi_woz_v22") NonMatchingChecksumError | Hi, I still meet the error when loading the datasets after upgradeing datasets.
raise NonMatchingChecksumError(error_msg + str(bad_urls))
datasets.utils.info_utils.NonMatchingChecksumError: Checksums didn't match for dataset source files:
['https://github.com/budzianowski/multiwoz/raw/master/data/MultiWOZ_2.2/dial... | Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumError(error_msg + str(bad_urls))
datasets.utils.info_utils.N... | 27 | load_dataset("multi_woz_v22") NonMatchingChecksumError
Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumErr... | [
-0.2244368196,
0.1764704585,
-0.0287197083,
0.1495724618,
0.1895724684,
0.0018985717,
0.3650580049,
0.4820297062,
0.2514896691,
0.166606307,
-0.0970347971,
0.1781591028,
-0.0822444707,
0.0992017165,
-0.184548825,
0.2732409537,
0.0765194371,
-0.0256412365,
-0.1682781428,
-0.0625... |
https://github.com/huggingface/datasets/issues/1876 | load_dataset("multi_woz_v22") NonMatchingChecksumError | This must be related to https://github.com/budzianowski/multiwoz/pull/72
Those files have changed, let me update the checksums for this dataset.
For now you can use `ignore_verifications=True` in `load_dataset` to skip the checksum verification. | Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumError(error_msg + str(bad_urls))
datasets.utils.info_utils.N... | 31 | load_dataset("multi_woz_v22") NonMatchingChecksumError
Hi, it seems that loading the multi_woz_v22 dataset gives a NonMatchingChecksumError.
To reproduce:
`dataset = load_dataset('multi_woz_v22','v2.2_active_only',split='train')`
This will give the following error:
```
raise NonMatchingChecksumErr... | [
-0.2244368196,
0.1764704585,
-0.0287197083,
0.1495724618,
0.1895724684,
0.0018985717,
0.3650580049,
0.4820297062,
0.2514896691,
0.166606307,
-0.0970347971,
0.1781591028,
-0.0822444707,
0.0992017165,
-0.184548825,
0.2732409537,
0.0765194371,
-0.0256412365,
-0.1682781428,
-0.0625... |
https://github.com/huggingface/datasets/issues/1872 | Adding a new column to the dataset after set_format was called | Hi ! Indeed if you add a column to a formatted dataset, then the new dataset gets a new formatting in which:
```
new formatted columns = (all columns - previously unformatted columns)
```
Therefore the new column is going to be formatted using the `torch` formatting.
If you want your new column to be unformatted... | Hi,
thanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side.
I load some lists of strings and integers, then call `data.set_format("torch", columns=["some_integer_column1"... | 67 | Adding a new column to the dataset after set_format was called
Hi,
thanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side.
I load some lists of strings and integers, th... | [
-0.1638377458,
-0.1748751998,
-0.0401732326,
-0.0492339544,
0.4729721546,
0.2625344396,
0.6921055913,
0.4177583754,
0.1687505543,
-0.2805864811,
0.1421068162,
0.3377909362,
-0.2340400815,
-0.0259004589,
0.1978134811,
-0.1026417166,
0.1268425435,
0.0638929904,
0.2461458296,
0.05... |
https://github.com/huggingface/datasets/issues/1872 | Adding a new column to the dataset after set_format was called | Ok cool :)
Also I just did a PR to mention this behavior in the documentation | Hi,
thanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side.
I load some lists of strings and integers, then call `data.set_format("torch", columns=["some_integer_column1"... | 16 | Adding a new column to the dataset after set_format was called
Hi,
thanks for the nice library. I'm in the process of creating a custom dataset, which has a mix of tensors and lists of strings. I stumbled upon an error and want to know if its a problem on my side.
I load some lists of strings and integers, th... | [
-0.1638377458,
-0.1748751998,
-0.0401732326,
-0.0492339544,
0.4729721546,
0.2625344396,
0.6921055913,
0.4177583754,
0.1687505543,
-0.2805864811,
0.1421068162,
0.3377909362,
-0.2340400815,
-0.0259004589,
0.1978134811,
-0.1026417166,
0.1268425435,
0.0638929904,
0.2461458296,
0.05... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | Hi @alejandrocros it looks like an incompatibility with the current Trainer @sgugger
Indeed currently the Trainer of `transformers` doesn't support a dataset with a transform
It looks like it comes from this line: https://github.com/huggingface/transformers/blob/f51188cbe74195c14c5b3e2e8f10c2f435f9751a/src/transfo... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 139 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | FYI that option can be removed with `remove_unused_columns = False` in your `TrainingArguments`, so there is a workaround @alexvaca0 while the fix in `Trainer` is underway.
@lhoestq I think I will just use the line you suggested and if someone is using the columns that are removed in their transform they will need t... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 75 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | I've tried your solutions @sgugger @lhoestq and the good news is that it throws no error. However, TPU training is taking forever, in 1 hour it has only trained 1 batch of 8192 elements, which doesn't make much sense... Is it possible that "on the fly" tokenization of batches is slowing down TPU training to that extent... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 57 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | I don't know what the value of `padding` is in your lines of code pasted above so I can't say for sure. The first batch will be very slow on TPU since it compiles everything, so that's normal (1 hour is long but 8192 elements is also large). Then if your batches are not of the same lengths, it will recompile everything... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 92 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | I have tried now on a GPU and it goes smooth! Amazing feature .set_transform() instead of .map()! Now I can pre-train my model without the hard disk limitation. Thanks for your work all HuggingFace team!! :clap: | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 36 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | In the end, to make it work I turned to A-100 gpus instead of TPUS, among other changes. Set_transform doesn't work as expected and slows down training very much even in GPUs, and applying map destroys the disk, as it multiplies by 100 the size of the data passed to it (due to inefficient implementation converting stri... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 179 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1867 | ERROR WHEN USING SET_TRANSFORM() | Great comment @alexvaca0 . I think that we could re-open the issue as a reformulation of why it takes so much space to save the arrow. Saving a 1% of oscar corpus takes more thank 600 GB (it breaks when it pass 600GB because it is the free memory that I have at this moment) when the full dataset is 1,3 TB. I have a 1TB... | Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__() missing 1 required positional arg... | 93 | ERROR WHEN USING SET_TRANSFORM()
Hi, I'm trying to use dataset.set_transform(encode) as @lhoestq told me in this issue: https://github.com/huggingface/datasets/issues/1825#issuecomment-774202797
However, when I try to use Trainer from transformers with such dataset, it throws an error:
```
TypeError: __init__... | [
-0.1826184243,
-0.2732515633,
0.1393029988,
0.1847400367,
0.7292690873,
0.1197054088,
0.6009588242,
0.2287761569,
-0.3227847219,
0.0544775538,
0.2355116159,
0.0333267935,
-0.0883407295,
-0.0234049894,
0.1892212331,
-0.5358201861,
-0.0138220638,
0.1288610548,
-0.4634346962,
-0.1... |
https://github.com/huggingface/datasets/issues/1859 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU) | Hi @corticalstack ! Thanks for reporting. Indeed in the recent versions of Faiss we must use `getDevice` to check if the index in on GPU.
I'm opening a PR | Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/impl/index_write.cpp:453: don't know how to serialize this type of index`
Note:
`torch.cuda.is_availabl... | 29 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU)
Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/... | [
-0.0712384656,
-0.3793424368,
-0.0126395077,
0.1242898777,
0.3250300586,
0.1440667957,
0.3014208376,
0.5358150601,
0.47758618,
0.3618178368,
0.0682532936,
-0.0320640914,
0.1739509106,
-0.0848500878,
-0.1260111034,
0.1587208807,
0.4708421528,
0.2794691622,
0.4284846187,
-0.04657... |
https://github.com/huggingface/datasets/issues/1859 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU) | I fixed this issue. It should work fine now.
Feel free to try it out by installing `datasets` from source.
Otherwise you can wait for the next release of `datasets` (in a few days) | Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/impl/index_write.cpp:453: don't know how to serialize this type of index`
Note:
`torch.cuda.is_availabl... | 34 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU)
Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/... | [
-0.2384791225,
-0.2405907959,
-0.0386404917,
0.1195870861,
0.3427580595,
0.1524924934,
0.3444892764,
0.5221388936,
0.4608198106,
0.2513040304,
-0.0791417286,
0.1412858516,
0.1216126382,
-0.05705918,
-0.1395934224,
0.1362418979,
0.4380635619,
0.1934212744,
0.301202327,
-0.058326... |
https://github.com/huggingface/datasets/issues/1859 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU) | Thanks for such a quick fix and merge to master, pip installed git master, tested all OK | Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/impl/index_write.cpp:453: don't know how to serialize this type of index`
Note:
`torch.cuda.is_availabl... | 17 | Error "in void don't know how to serialize this type of index" when saving index to disk when device=0 (GPU)
Error serializing faiss index. Error as follows:
`Error in void faiss::write_index(const faiss::Index*, faiss::IOWriter*) at /home/conda/feedstock_root/build_artifacts/faiss-split_1612472484670/work/faiss/... | [
-0.1146761328,
-0.3390221298,
-0.0211909842,
0.1105612442,
0.3648000658,
0.1579404324,
0.2933630049,
0.5287505984,
0.5046113729,
0.3331743479,
0.0193730295,
0.0713493749,
0.111765027,
-0.0630021468,
-0.1396044493,
0.1634269953,
0.4838933945,
0.2276031077,
0.3357010186,
-0.03815... |
https://github.com/huggingface/datasets/issues/1857 | Unable to upload "community provided" dataset - 400 Client Error | Hi ! We're in the process of switching the community datasets to git repos, exactly like what we're doing for models.
You can find an example here:
https://huggingface.co/datasets/lhoestq/custom_squad/tree/main
We'll update the CLI in the coming days and do a new release :)
Also cc @julien-c maybe we can make i... | Hi,
i'm trying to a upload a dataset as described [here](https://huggingface.co/docs/datasets/v1.2.0/share_dataset.html#sharing-a-community-provided-dataset). This is what happens:
```
$ datasets-cli login
$ datasets-cli upload_dataset my_dataset
About to upload file /path/to/my_dataset/dataset_infos.json to S3... | 54 | Unable to upload "community provided" dataset - 400 Client Error
Hi,
i'm trying to a upload a dataset as described [here](https://huggingface.co/docs/datasets/v1.2.0/share_dataset.html#sharing-a-community-provided-dataset). This is what happens:
```
$ datasets-cli login
$ datasets-cli upload_dataset my_dataset... | [
-0.091570951,
-0.0363661684,
0.0344167911,
0.0672418997,
0.3649701476,
-0.0492852964,
0.1734050065,
0.0044842041,
-0.2230457217,
-0.1468115449,
-0.1488375813,
-0.0205999389,
0.0050856299,
0.3581643999,
-0.0571106076,
-0.106930241,
-0.0269480255,
0.0025080862,
-0.4462412,
0.0053... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | Hi ! This issue may be related to #996
This comes probably from the Quota Exceeded error from Google Drive.
Can you try again tomorrow and see if you still have the error ?
On my side I didn't get any error today with `load_dataset("amazon_polarity")` | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 45 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.1446352154,
0.1181075647,
-0.1213069111,
0.241092965,
0.1280619651,
-0.0035924024,
0.3443529308,
0.0652999952,
0.2865490913,
0.2393917739,
0.0281221364,
-0.0647890642,
0.0097156595,
0.150679931,
-0.1385418922,
0.2059793323,
0.0107033998,
-0.0445431918,
-0.1888068914,
0.17735... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | @lhoestq Hi! I encounter the same error when loading `yelp_review_full`.
```
from datasets import load_dataset
dataset_yp = load_dataset("yelp_review_full")
```
When you say the "Quota Exceeded from Google drive". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive? | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 45 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.0611734912,
0.1423272789,
-0.1086320728,
0.2213834375,
0.1263364702,
0.1217507794,
0.2452766448,
0.0103801079,
0.3931342959,
0.1096664593,
-0.1395590156,
-0.0201450661,
0.0363820083,
0.025237916,
-0.1712937504,
0.36432302,
0.0185450204,
-0.0113466484,
-0.0788384005,
-0.01027... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | > When you say the "Quota Exceeded from Google drive". Is this a quota from the dataset owner? or the quota from our (the runner) Google Drive?
Each file on Google Drive can be downloaded only a certain amount of times per day because of a quota. The quota is reset every day. So if too many people download the datas... | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 127 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.1349426359,
0.0474776924,
-0.1110619456,
0.2588098645,
0.0624817871,
-0.0367825255,
0.2863383889,
0.0119939027,
0.3435027599,
0.2581291199,
0.0767342001,
-0.0375791937,
0.0030820393,
0.0379521847,
-0.233708486,
0.1695693135,
-0.0217132568,
-0.0736347213,
-0.1117963046,
0.040... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | @lhoestq Gotcha, that is quite problematic...for what it's worth, I've had no issues with the other datasets I tried, such as `yelp_reviews_full` and `amazon_reviews_multi`. | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 24 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.1947867572,
0.0554472394,
-0.0836697891,
0.1578220427,
0.1719483733,
0.0804431066,
0.2375734597,
0.0991518348,
0.2416200191,
-0.0136138508,
-0.083981894,
0.0607006066,
-0.0039982293,
-0.1410386264,
-0.3183495998,
0.3028694093,
0.0808663219,
0.0319711976,
-0.1793726832,
0.036... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | Same issue today with "big_patent", though the symptoms are slightly different.
When running
```py
from datasets import load_dataset
load_dataset("big_patent", split="validation")
```
I get the following
`FileNotFoundError: Local file \huggingface\datasets\downloads\6159313604f4f2c01e7d1cac52139343b6c07f73... | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 230 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.1977479905,
0.1699574441,
-0.0510926507,
0.2398214042,
0.1494402289,
-0.0500103757,
0.2648404837,
0.2264508009,
0.318641901,
0.0602950603,
-0.0621271282,
0.0072554699,
-0.0048484444,
-0.1367177367,
-0.2075067312,
0.3518306315,
0.1385654658,
-0.0169534143,
-0.2259688377,
0.01... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | A similar issue arises when trying to stream the dataset
```python
>>> from datasets import load_dataset
>>> iter_dset = load_dataset("amazon_polarity", split="test", streaming=True)
>>> iter(iter_dset).__next__()
---------------------------------------------------------------------------
ValueError ... | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 430 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.3396336734,
0.0200679544,
-0.0462277047,
0.1762584299,
0.1373029798,
0.0859515518,
0.199638918,
0.1355160624,
0.2324956357,
0.0852484256,
-0.0440213904,
0.0785093829,
-0.0269583929,
-0.0893351138,
-0.2416516542,
0.1742836535,
0.0135003915,
0.078339465,
-0.181372717,
0.080216... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | This error still happens, but for a different reason now: Google Drive returns a warning instead of the dataset. | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 19 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.2818935215,
0.1576435864,
-0.0683750734,
0.0996028557,
0.108074218,
0.1604389548,
0.4640991092,
0.0867969096,
0.2902540565,
0.1448861808,
0.0573569424,
-0.0298759751,
-0.1013723463,
0.0111130737,
-0.2888023853,
0.2514447272,
0.1300302893,
0.0149664385,
-0.0264263805,
0.02532... |
https://github.com/huggingface/datasets/issues/1856 | load_dataset("amazon_polarity") NonMatchingChecksumError | Hi ! Thanks for reporting. Google Drive changed the way to bypass the warning message recently.
The latest release `1.18.4` fixes this for datasets loaded in a regular way.
We opened a PR to fix this recently for streaming mode at #3843 - we'll do a new release once the fix is merged :) | Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
---------------------------------------------------------------------------
NonMatchingChecksumError Traceback ... | 54 | load_dataset("amazon_polarity") NonMatchingChecksumError
Hi, it seems that loading the amazon_polarity dataset gives a NonMatchingChecksumError.
To reproduce:
```
load_dataset("amazon_polarity")
```
This will give the following error:
```
----------------------------------------------------------------------... | [
-0.3437190056,
0.0966544077,
-0.0553667955,
0.0980792046,
0.1595421582,
0.1303638965,
0.4060623348,
0.1062448397,
0.2796753943,
0.1851273179,
0.0166730359,
-0.0457288623,
-0.0545699187,
0.079530865,
-0.3115005791,
0.1836310774,
0.0936344117,
0.0024180119,
-0.0331716463,
0.05033... |
https://github.com/huggingface/datasets/issues/1854 | Feature Request: Dataset.add_item | Hi @sshleifer.
I am not sure of understanding the need of the `add_item` approach...
By just reading your "Desired API" section, I would say you could (nearly) get it with a 1-column Dataset:
```python
data = {"input_ids": [np.array([4,4,2]), np.array([8,6,5,5,2]), np.array([3,3,31,5])]}
ds = Dataset.from_dict... | I'm trying to integrate `huggingface/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.m... | 48 | Feature Request: Dataset.add_item
I'm trying to integrate `huggingface/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L318, as opposed to loading all the ... | [
-0.2646300197,
0.1603153199,
-0.0458906069,
0.1386156231,
-0.0252808258,
0.1201699898,
0.1569379121,
0.1071406603,
0.0586702526,
0.0440865979,
0.1095490381,
0.5823286176,
-0.1573916227,
0.1380069703,
0.445127517,
-0.1356468052,
0.1021328643,
0.2048072666,
-0.1752962321,
0.07040... |
https://github.com/huggingface/datasets/issues/1854 | Feature Request: Dataset.add_item | Hi @sshleifer :)
We don't have methods like `Dataset.add_batch` or `Dataset.add_entry/add_item` yet.
But that's something we'll add pretty soon. Would an API that looks roughly like this help ? Do you have suggestions ?
```python
import numpy as np
from datasets import Dataset
tokenized = [np.array([4,4,2]),... | I'm trying to integrate `huggingface/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L318, as opposed to loading all the text into arrow, and then `dataset.m... | 92 | Feature Request: Dataset.add_item
I'm trying to integrate `huggingface/datasets` functionality into `fairseq`, which requires (afaict) being able to build a dataset through an `add_item` method, such as https://github.com/pytorch/fairseq/blob/master/fairseq/data/indexed_dataset.py#L318, as opposed to loading all the ... | [
-0.2646300197,
0.1603153199,
-0.0458906069,
0.1386156231,
-0.0252808258,
0.1201699898,
0.1569379121,
0.1071406603,
0.0586702526,
0.0440865979,
0.1095490381,
0.5823286176,
-0.1573916227,
0.1380069703,
0.445127517,
-0.1356468052,
0.1021328643,
0.2048072666,
-0.1752962321,
0.07040... |
https://github.com/huggingface/datasets/issues/1849 | Add TIMIT | @patrickvonplaten Could you please help me with how the output text has to be represented in the data? TIMIT has Words, Phonemes and texts. Also has lot on info on the speaker and the dialect. Could you please help me? An example of how to arrange it would be super helpful!
| ## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.inf.ed.ac.uk... | 51 | Add TIMIT
## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.... | [
0.073341012,
-0.311024338,
-0.1762830168,
0.1887155622,
0.0858768448,
-0.1072418094,
0.1176899076,
0.0632868633,
-0.4261907339,
-0.135892272,
-0.2996530831,
0.1368212998,
-0.0700435638,
0.3138228953,
0.2315153331,
-0.3812041581,
-0.0924159214,
0.1239683926,
-0.2485863864,
-0.16... |
https://github.com/huggingface/datasets/issues/1849 | Add TIMIT | Hey @vrindaprabhu - sure I'll help you :-) Could you open a first PR for TIMIT where you copy-paste more or less the `librispeech_asr` script: https://github.com/huggingface/datasets/blob/28be129db862ec89a87ac9349c64df6b6118aff4/datasets/librispeech_asr/librispeech_asr.py#L93 (obviously replacing all the naming and lin... | ## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.inf.ed.ac.uk... | 85 | Add TIMIT
## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.... | [
-0.2446836978,
-0.4731052816,
-0.1747952849,
0.0511884764,
0.1257332265,
-0.090471305,
0.0721050799,
0.2155633718,
-0.2834858298,
0.1218879744,
-0.302002579,
0.1912913769,
-0.3662652671,
0.1645170301,
0.1355875283,
-0.0494575873,
-0.1529121697,
0.1166827679,
-0.0651813149,
-0.1... |
https://github.com/huggingface/datasets/issues/1849 | Add TIMIT | I am sorry! I created the PR [#1903](https://github.com/huggingface/datasets/pull/1903#). Requesting your comments! CircleCI tests are failing, will address them along with your comments! | ## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.inf.ed.ac.uk... | 22 | Add TIMIT
## Adding a Dataset
- **Name:** *TIMIT*
- **Description:** *The TIMIT corpus of read speech has been designed to provide speech data for the acquisition of acoustic-phonetic knowledge and for the development and evaluation of automatic speech recognition systems*
- **Paper:** *Homepage*: http://groups.... | [
-0.2418492734,
-0.3422918618,
-0.1415261924,
0.0710315555,
-0.0340814777,
-0.0950556919,
0.1323645115,
0.1322911233,
-0.3182581067,
0.2184931189,
-0.3416488171,
0.1853345186,
-0.2274642438,
0.2172826678,
0.1103417352,
-0.2743068933,
-0.0984167755,
0.0761338696,
0.0409775339,
-0... |
https://github.com/huggingface/datasets/issues/1844 | Update Open Subtitles corpus with original sentence IDs | Hi ! You're right this can can useful.
This should be easy to add, so feel free to give it a try if you want to contribute :)
I think we just need to add it to the _generate_examples method of the OpenSubtitles dataset builder [here](https://github.com/huggingface/datasets/blob/master/datasets/open_subtitles/open_sub... | Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat a... | 46 | Update Open Subtitles corpus with original sentence IDs
Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media i... | [
0.2011131197,
0.1604523957,
-0.0483416542,
-0.2089408487,
-0.1301653534,
0.2428105772,
0.2711244524,
0.2571243346,
-0.2969916463,
-0.038673833,
-0.31456393,
0.2453639507,
0.2158561498,
-0.0763090327,
0.0854991004,
-0.332557112,
0.1179938167,
0.0025806008,
0.5615066886,
-0.20760... |
https://github.com/huggingface/datasets/issues/1844 | Update Open Subtitles corpus with original sentence IDs | Hey @lhoestq , absolutely yes! Just one question before I start implementing. The ids found in the zip file have this format:
(the following is line `22497315` of the `ids` file of the `de-en` dump)
`de/2017/7006210/7063319.xml.gz en/2017/7006210/7050201.xml.gz 335 339 340` (every space is actually a tab, ... | Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat a... | 217 | Update Open Subtitles corpus with original sentence IDs
Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media i... | [
0.2530397177,
0.248258397,
-0.0133867636,
-0.1263540089,
-0.2649044096,
0.1654840112,
0.2578909695,
0.2178696841,
-0.3385523558,
-0.1133100837,
-0.3191738427,
0.3215388954,
0.1591857821,
-0.1485190392,
0.0773391724,
-0.2948979139,
0.1316312701,
-0.1250015348,
0.4648746848,
-0.2... |
https://github.com/huggingface/datasets/issues/1844 | Update Open Subtitles corpus with original sentence IDs | I like the idea of having `year`, `imdbId` and `subtitleId` as columns for filtering for example.
And for the `sentenceIds` a list of integers is fine. | Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat a... | 26 | Update Open Subtitles corpus with original sentence IDs
Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media i... | [
0.2025861889,
0.152048111,
-0.0463983528,
-0.2151802629,
-0.2097587734,
0.2218637168,
0.2927141786,
0.3270961046,
-0.280744791,
-0.0863662586,
-0.3423936665,
0.1748221964,
0.1811572462,
-0.1334892213,
0.0285321381,
-0.3397962153,
0.1369059682,
-0.0015651578,
0.6896281838,
-0.22... |
https://github.com/huggingface/datasets/issues/1844 | Update Open Subtitles corpus with original sentence IDs | Something like this? (adapted from [here](https://github.com/huggingface/datasets/blob/master/datasets/open_subtitles/open_subtitles.py#L114))
```python
result = (
sentence_counter,
{
"id": str(sentence_counter),
"meta": {
"year": year,
"imdbId": imdb_id,
"subtitleId... | Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media id, subtitle file id and sentence id), therefore somewhat a... | 79 | Update Open Subtitles corpus with original sentence IDs
Hi! It would be great if you could add the original sentence ids to [Open Subtitles](https://huggingface.co/datasets/open_subtitles).
I can think of two reasons: first, it's possible to gather sentences for an entire document (the original ids contain media i... | [
0.2209783942,
0.0347742066,
-0.0174860228,
-0.1517955363,
-0.2049597502,
0.3104093075,
0.3499245048,
0.2389915138,
-0.4182049036,
-0.1190219373,
-0.3992201388,
0.2717108727,
0.2253196388,
-0.1207072362,
0.1057459861,
-0.2710323632,
0.1577948779,
-0.0647952259,
0.5990245342,
-0.... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | That's awesome! Actually, I just noticed that this dataset might become a bit too big!
MuST-C is the main dataset used for IWSLT19 and should probably be added as a standalone dataset. Would you be interested also in adding `datasets/MuST-C` instead?
Description:
_MuST-C is a multilingual speech translation cor... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 188 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.2875131071,
0.1752748936,
-0.052337721,
-0.0160809606,
-0.1027402133,
0.0413462855,
0.0322459415,
0.1573133767,
-0.3544580936,
0.2396766543,
-0.2381985486,
-0.111177206,
-0.1982450485,
0.1046884805,
0.1299679875,
-0.1173932999,
0.0626323596,
0.162166208,
-0.2575621009,
-0.24... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Hi @patrickvonplaten
I have tried downloading this dataset, but the connection seems to reset all the time. I have tried it via the browser, wget, and using gdown . But it gives me an error message. _"The server is busy or down, pls try again"_ (rephrasing the message here)
I have completed adding 4 datasets in th... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 90 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.4057503343,
0.1195164621,
-0.0039104498,
0.1509026587,
0.0220462717,
-0.0853216723,
-0.1335305572,
-0.0062556555,
-0.240002498,
0.1257293224,
-0.2750510275,
-0.3189472556,
0.1023330241,
0.1398472637,
0.1449169517,
-0.093930915,
0.1205428839,
0.0259056035,
-0.3361466825,
-0.2... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | @skyprince999, I think I'm getting the same error you're getting :-/
```
Sorry, you can't view or download this file at this time.
Too many users have viewed or downloaded this file recently. Please try accessing the file again later. If the file you are trying to access is particularly large or is shared with m... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 117 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.3741523623,
0.0519903637,
-0.0482890569,
0.0795519277,
-0.0135706309,
0.0174990743,
0.0815657154,
0.1886134893,
-0.2064035386,
0.3855750263,
-0.3391421139,
-0.5373646021,
-0.0532581173,
-0.0104366131,
0.1305615604,
-0.2101035863,
0.0005455129,
-0.0619699098,
-0.2461218834,
-... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Also there are huge those datasets. Think downloading MuST-C v1.2 amounts to ~ 1000GB... because there are 14 possible configs each around 60-70GB. I think users mostly will only use one of the 14 configs so that they would only need, in theory, will have to download ~60GB which is ok. But I think this functionality do... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 64 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.3945274949,
0.2756193876,
-0.0875780582,
0.0062158853,
-0.0652419478,
0.0286357198,
-0.1824723333,
0.2685896158,
-0.2595131397,
0.3226704001,
-0.2936765254,
-0.2957717478,
-0.1945613623,
0.1870193183,
0.2089398205,
0.0910525322,
0.0300667007,
0.2597108185,
-0.4889111519,
-0.... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | > Also cc @lhoestq - do you think we could mirror the dataset?
Yes we can mirror it if the authors are fine with it. You can create a dataset repo on huggingface.co (possibly under the relevant org) and add the mirrored data files.
> I think users mostly will only use one of the 14 configs so that they would only... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 110 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.3701687157,
0.0756935403,
-0.0500219986,
0.1428011358,
-0.1143721044,
-0.0225354824,
0.0245043971,
0.252882123,
-0.2171292156,
0.2163131684,
-0.3634769917,
-0.2713862658,
-0.095288828,
0.2548370361,
0.1178232208,
-0.225666374,
0.0715245306,
0.062893115,
-0.4934399128,
-0.180... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | I have written to the dataset authors, highlighting this issue. Waiting for their response.
Update on 25th Feb:
The authors have replied back, they are updating the download link and will revert back shortly!
```
first of all thanks a lot for being interested in MuST-C and for building the data-loader.
Be... | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 147 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.199468717,
0.243653208,
-0.0974726528,
-0.0784686431,
-0.0106336009,
-0.145165652,
0.0524758771,
0.075011231,
-0.1210071668,
0.3005027771,
-0.1688124239,
-0.121463716,
-0.0929061249,
0.1014920101,
0.0492404997,
0.0520472825,
0.0159371328,
0.1417452991,
-0.4235428572,
-0.2137... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Awesome, actually @lhoestq let's just ask the authors if we should host the dataset no? They could just use our links then as well for their website - what do you think? Is it fine to use our AWS dataset storage also as external links? | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 45 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.1768889874,
0.2350076586,
-0.0860685408,
0.2041038275,
-0.2330979258,
-0.0434643999,
0.3742260933,
-0.0443167314,
0.065488033,
0.1011237428,
-0.4423069358,
-0.2948478758,
-0.0530128293,
0.190222472,
0.1885555983,
0.0627990291,
0.1907826811,
-0.2238711566,
-0.1670014411,
-0.1... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Yes definitely. Shall we suggest them to create a dataset repository under their org on huggingface.co ? @julien-c
The dataset is around 1TB | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 23 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.2819740176,
0.1316505522,
-0.1142895445,
0.1891250461,
-0.0761728585,
0.0238954872,
-0.0151866302,
0.1323354244,
-0.2927866876,
0.1568828225,
-0.2882781029,
-0.2273532599,
-0.1356900185,
0.177146107,
0.1977109164,
-0.1647575796,
0.1265197098,
-0.0231165066,
-0.32716766,
-0.2... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Sounds good!
Order of magnitude is storage costs ~$20 per TB per month (not including bandwidth).
Happy to provide this to the community as I feel this is an important dataset. Let us know what the authors want to do!
| ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 40 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.2674562335,
0.346837908,
-0.1150630787,
0.0743558854,
-0.1094926745,
0.0280107968,
0.0628128126,
0.2007726729,
-0.3066084683,
0.218983233,
-0.2691019475,
-0.2065333873,
-0.1899873465,
0.1393337548,
0.1780731678,
-0.1505706161,
0.1918691993,
0.0074417484,
-0.287493825,
-0.270... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Great! @skyprince999, do you think you could ping the authors here or link to this thread? I think it could be a cool idea to host the dataset on our side then | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 32 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.2638153136,
0.2556505799,
-0.1020891815,
0.0674262643,
-0.1163010672,
0.004987149,
0.2088762373,
0.1324666291,
-0.2380465567,
0.1958874315,
-0.2925174832,
-0.194789961,
-0.1235652119,
0.1762094349,
0.2179547399,
-0.0965975821,
0.1338677108,
-0.0480609424,
-0.2836481929,
-0.3... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Done. They replied back, and they want to have a call over a meet/ skype. Is that possible ?
Btw @patrickvonplaten you are looped in that email (_pls check you gmail account_) | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 32 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.2442643195,
0.076314114,
-0.1269343644,
0.129224956,
-0.0602060109,
-0.108165808,
0.0954099,
0.0292593278,
-0.1983939409,
0.2287622988,
-0.2163419574,
-0.3068775535,
-0.0361022353,
0.2048379481,
0.1446813196,
-0.1205032244,
-0.0763828158,
0.0352110974,
-0.2530812919,
-0.3930... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | @gegallego there were some concerns regarding dataset usage & attribution by a for-profit company, so couldn't take it forward. Also the download links were unstable.
But I guess if you want to test the fairseq benchmarks, you can connect with them directly for downloading the dataset. | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 46 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.3551786244,
0.273317337,
-0.0848914683,
0.0439707823,
-0.2106074244,
-0.0997216329,
0.0143610621,
0.1664300412,
-0.3334970176,
0.1858730912,
-0.2543599606,
-0.0773929209,
0.0546254516,
-0.1191614121,
0.1883631051,
-0.2674563527,
0.1272602826,
0.0159448534,
-0.4990531206,
-0.... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | Yes, that dataset is not easy to download... I had to copy it to my Google Drive and use `rsync` to be able to download it.
However, we could add the dataset with a manual download, right? | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 37 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.3261700869,
0.2585772276,
-0.0921743289,
-0.0333682485,
-0.0554309823,
0.0201776382,
0.0042866687,
0.1221485808,
-0.33253479,
0.2870287597,
-0.2463111728,
-0.1663524061,
-0.0883381367,
0.0430672206,
0.1519542634,
-0.0460928828,
0.1005881503,
0.0084113041,
-0.2816007733,
-0.3... |
https://github.com/huggingface/datasets/issues/1843 | MustC Speech Translation | yes that is possible. I couldn't unfortunately complete this PR, If you would like to add it, please feel free to do it. | ## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.com/view/iwslt-evaluation-2... | 23 | MustC Speech Translation
## Adding a Dataset
- **Name:** *IWSLT19*
- **Description:** *The Speech Translation Task addresses the translation of English audio into German and Portuguese text.*
- **Hompage:** *https://sites.google.com/view/iwslt-evaluation-2019/speech-translation*
- **Data:** *https://sites.google.... | [
-0.314416647,
0.2407134175,
-0.0985174626,
0.0024262667,
-0.0414731912,
-0.0363332331,
0.0861026645,
0.156312719,
-0.3312878311,
0.1687451452,
-0.3144814372,
-0.0872286037,
-0.080084078,
0.1177136153,
0.2429231256,
-0.2341310978,
0.1259044558,
0.0504560396,
-0.3424839973,
-0.24... |
https://github.com/huggingface/datasets/issues/1840 | Add common voice | Hey @BirgerMoell - awesome that you started working on Common Voice. Common Voice is a bit special since, there is no direct download link to download the data. In these cases we usually consider two options:
1) Find a hacky solution to extract the download link somehow from the XLM tree of the website
2) If this ... | ## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www.tensorflow.org/dat... | 100 | Add common voice
## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www... | [
-0.1914141923,
-0.2330224067,
-0.0579187907,
-0.1640419811,
0.1464049071,
-0.077341564,
0.2238349319,
0.2807313204,
-0.2090620697,
0.3071741462,
-0.4487543404,
0.0414791219,
-0.0504973456,
-0.0618108138,
0.1337497085,
-0.1799684763,
-0.2989138961,
0.0928418636,
0.0731950849,
-0... |
https://github.com/huggingface/datasets/issues/1840 | Add common voice | I added a Work in Progress pull request (hope that is ok). I've made a card for the dataset and filled out the common_voice.py file with information about the datset (not completely).
I didn't manage to get the tagging tool working locally on my machine but will look into that later.
Left to do.
- Tag the data... | ## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www.tensorflow.org/dat... | 66 | Add common voice
## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www... | [
-0.1344469786,
-0.2228565216,
0.0001930711,
-0.1578858197,
0.1937068105,
0.004948426,
0.2681258917,
0.3020457923,
-0.2495007366,
0.1795227081,
-0.3955019414,
0.2300344557,
-0.0425767787,
-0.0264863092,
0.0067094136,
-0.1008290797,
-0.1025350168,
0.1595254391,
0.1041881964,
-0.2... |
https://github.com/huggingface/datasets/issues/1840 | Add common voice | I saw that this current datasets package holds common voice version 6.1, how to add the new version 7.0 that is already available? | ## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www.tensorflow.org/dat... | 23 | Add common voice
## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www... | [
-0.3583086133,
-0.3505197465,
-0.0120856063,
-0.2336004674,
-0.4511424005,
0.0678055808,
-0.0332449079,
0.54031986,
-0.1008659303,
0.0819962397,
-0.2226713747,
0.246444717,
-0.2389799803,
0.062518917,
-0.0957735404,
-0.2583046556,
0.1143465862,
0.1522572637,
0.1945268363,
-0.17... |
https://github.com/huggingface/datasets/issues/1840 | Add common voice | Common voice still appears to be a 6.1. Is the plan still to upgrade to 7.0? | ## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www.tensorflow.org/dat... | 16 | Add common voice
## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www... | [
-0.21228908,
-0.3657412827,
-0.0668141395,
-0.2508972585,
-0.1655759215,
-0.1441819966,
0.2083906829,
0.5740472674,
-0.1634231508,
0.1866548359,
-0.1267237514,
0.1299787313,
-0.2213794142,
-0.0540739633,
-0.1652476043,
-0.172515437,
0.1874604821,
0.0923388228,
0.2241402566,
-0.... |
https://github.com/huggingface/datasets/issues/1840 | Add common voice | We actually already have the code and everything ready to add Common Voice 7.0 to `datasets` but are still waiting for the common voice authors to give us the green light :-)
Also gently pinging @phirework and @milupo here | ## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www.tensorflow.org/dat... | 39 | Add common voice
## Adding a Dataset
- **Name:** *common voice*
- **Description:** *Mozilla Common Voice Dataset*
- **Paper:** Homepage: https://voice.mozilla.org/en/datasets
- **Data:** https://voice.mozilla.org/en/datasets
- **Motivation:** Important speech dataset
- **TFDatasets Implementation**: https://www... | [
-0.2997157574,
-0.3046939969,
-0.0558255352,
-0.2554713488,
-0.045264665,
-0.1200117841,
0.3751796186,
0.4150421321,
-0.1786259264,
0.2442779988,
-0.2419281453,
0.2421715707,
-0.1680073142,
0.0401818268,
-0.1121170595,
-0.2051170021,
-0.1020666957,
0.1860423386,
0.1300929189,
-... |
https://github.com/huggingface/datasets/issues/1838 | Add tedlium | Hi @patrickvonplaten
I can have a look to this dataset later since I am trying to add the OpenSLR dataset https://github.com/huggingface/datasets/pull/2173
Hopefully I have enough space since the compressed file is 21GB. The release 3 is even bigger: 54GB :-0 | ## Adding a Dataset
- **Name:** *tedlium*
- **Description:** *The TED-LIUM 1-3 corpus is English-language TED talks, with transcriptions, sampled at 16kHz. It contains about 118 hours of speech.*
- **Paper:** Homepage: http://www.openslr.org/7/, https://lium.univ-lemans.fr/en/ted-lium2/ &, https://www.openslr.org/51... | 40 | Add tedlium
## Adding a Dataset
- **Name:** *tedlium*
- **Description:** *The TED-LIUM 1-3 corpus is English-language TED talks, with transcriptions, sampled at 16kHz. It contains about 118 hours of speech.*
- **Paper:** Homepage: http://www.openslr.org/7/, https://lium.univ-lemans.fr/en/ted-lium2/ &, https://www.... | [
-0.343177408,
0.0705062822,
-0.0683077797,
0.1333119571,
0.0683672652,
-0.0204196069,
0.0353582837,
0.3795070052,
-0.4294745326,
0.2744410932,
-0.2579483092,
0.3504007757,
-0.208436802,
-0.0917881653,
0.2269741446,
-0.3618194759,
-0.1412835568,
0.1973716766,
-0.1658692956,
-0.1... |
https://github.com/huggingface/datasets/issues/1837 | Add VCTK | @patrickvonplaten I'd like to take this, if nobody has already done it. I have added datasets before through the datasets sprint, but I feel rusty on the details, so I'll look at the guide as well as similar audio PRs (#1878 in particular comes to mind). If there is any detail I should be aware of please, let me know! ... | ## Adding a Dataset
- **Name:** *VCTK*
- **Description:** *This CSTR VCTK Corpus includes speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent arch... | 72 | Add VCTK
## Adding a Dataset
- **Name:** *VCTK*
- **Description:** *This CSTR VCTK Corpus includes speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech ... | [
-0.3240322471,
-0.161468178,
-0.0846254602,
-0.1376853883,
0.013386514,
-0.164779678,
0.299248904,
0.1571182609,
-0.3928883672,
0.3696943223,
-0.1535666138,
0.2089087963,
-0.1626086384,
0.0988039225,
0.1848953664,
-0.4437522292,
-0.0921772271,
0.1608295739,
0.0118198572,
-0.156... |
https://github.com/huggingface/datasets/issues/1837 | Add VCTK | That sounds great @jaketae - let me know if you need any help i.e. feel free to ping me on a first PR :-) | ## Adding a Dataset
- **Name:** *VCTK*
- **Description:** *This CSTR VCTK Corpus includes speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech accent arch... | 24 | Add VCTK
## Adding a Dataset
- **Name:** *VCTK*
- **Description:** *This CSTR VCTK Corpus includes speech data uttered by 110 English speakers with various accents. Each speaker reads out about 400 sentences, which were selected from a newspaper, the rainbow passage and an elicitation paragraph used for the speech ... | [
-0.3135918975,
-0.2029017955,
-0.1356343627,
-0.0900804475,
-0.0944114998,
-0.1734262109,
0.3951478601,
0.2453563958,
-0.3127599955,
0.2707165778,
-0.1096361205,
0.1168668866,
-0.1142661646,
0.130008176,
0.1152627692,
-0.3554119468,
-0.1647337675,
0.1606166512,
-0.1079571769,
-... |
https://github.com/huggingface/datasets/issues/1831 | Some question about raw dataset download info in the project . | Hi ! The `dl_manager` is a `DownloadManager` object and is responsible for downloading the raw data files.
It is used by dataset builders in their `_split_generators` method to download the raw data files that are necessary to build the datasets splits.
The `Conll2003` class is a dataset builder, and so you can dow... | Hi , i review the code in
https://github.com/huggingface/datasets/blob/master/datasets/conll2003/conll2003.py
in the _split_generators function is the truly logic of download raw datasets with dl_manager
and use Conll2003 cls by use import_main_class in load_dataset function
My question is that , with this logic i... | 166 | Some question about raw dataset download info in the project .
Hi , i review the code in
https://github.com/huggingface/datasets/blob/master/datasets/conll2003/conll2003.py
in the _split_generators function is the truly logic of download raw datasets with dl_manager
and use Conll2003 cls by use import_main_class ... | [
-0.1392953396,
-0.0969951972,
-0.0575437993,
0.531555295,
0.1806543171,
-0.0776940212,
0.1291474253,
-0.0684931949,
0.13074781,
0.1263671964,
-0.4010900557,
0.190399006,
-0.0717180073,
0.5283398032,
0.1642403007,
-0.0650983974,
-0.1106834039,
0.1475781649,
-0.0933765322,
-0.073... |
https://github.com/huggingface/datasets/issues/1831 | Some question about raw dataset download info in the project . | I am afraid that there is not a very straightforward way to get that location.
Another option, from _split_generators would be to use:
- `dl_manager._download_config.cache_dir` to get the directory where all the raw downloaded files are:
```python
download_dir = dl_manager._download_config.cache_dir
```
-... | Hi , i review the code in
https://github.com/huggingface/datasets/blob/master/datasets/conll2003/conll2003.py
in the _split_generators function is the truly logic of download raw datasets with dl_manager
and use Conll2003 cls by use import_main_class in load_dataset function
My question is that , with this logic i... | 111 | Some question about raw dataset download info in the project .
Hi , i review the code in
https://github.com/huggingface/datasets/blob/master/datasets/conll2003/conll2003.py
in the _split_generators function is the truly logic of download raw datasets with dl_manager
and use Conll2003 cls by use import_main_class ... | [
-0.1346984059,
-0.029141767,
-0.0565955229,
0.5307530761,
0.1188368648,
-0.1144692823,
0.1020082533,
-0.0082433783,
0.0415086821,
0.0833945274,
-0.419790566,
0.2072835416,
-0.1245642379,
0.5449793339,
0.1708202511,
-0.0960756242,
-0.0993764848,
0.1221579462,
-0.1015596911,
-0.0... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.